sandy bridge 2700k 2021

Intel Core i7-2700K testing with a BIOSTAR B75MU3B v5.0 (4.6.5 BIOS) and Intel HD 3000 2GB on Ubuntu 21.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2106203-IB-SANDYBRID51
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results

Limit displaying results to tests within:

AV1 3 Tests
Timed Code Compilation 3 Tests
C/C++ Compiler Tests 6 Tests
CPU Massive 7 Tests
Creator Workloads 7 Tests
Encoding 5 Tests
HPC - High Performance Computing 4 Tests
Machine Learning 3 Tests
Multi-Core 10 Tests
Programmer / Developer System Benchmarks 3 Tests
Server CPU Tests 5 Tests
Video Encoding 5 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
1
June 18 2021
  1 Day, 6 Hours, 44 Minutes
2
June 19 2021
  10 Hours, 22 Minutes
3
June 19 2021
  10 Hours, 23 Minutes
Invert Hiding All Results Option
  17 Hours, 10 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


sandy bridge 2700k 2021ProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerOpenGLVulkanCompilerFile-SystemScreen Resolution123Intel Core i7-2700K @ 3.90GHz (4 Cores / 8 Threads)BIOSTAR B75MU3B v5.0 (4.6.5 BIOS)Intel 2nd Generation Core DRAM8GB525GB Crucial CT525MX3Intel HD 3000 2GB (1350MHz)Realtek ALC662 rev1G237HLRealtek RTL8111/8168/8411Ubuntu 21.045.11.0-18-generic (x86_64)GNOME Shell 3.38.4X Server + Wayland3.3 Mesa 21.2.0-devel (git-849ab4e 2021-06-17 hirsute-oibaf-ppa)1.1.180GCC 10.3.0ext41920x1080OpenBenchmarking.orgKernel Details- Transparent Huge Pages: madviseCompiler Details- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-mutex --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-10-gDeRY6/gcc-10-10.3.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-10-gDeRY6/gcc-10-10.3.0/debian/tmp-gcn/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- Scaling Governor: intel_cpufreq schedutil - CPU Microcode: 0x2f - Thermald 2.4.3Security Details- itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Mitigation of PTE Inversion; VMX: conditional cache flushes SMT vulnerable + mds: Mitigation of Clear buffers; SMT vulnerable + meltdown: Mitigation of PTI + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected

123Result OverviewPhoronix Test Suite100%100%101%101%101%Timed GDB GNU Debugger CompilationSVT-AV1Timed FFmpeg CompilationNCNNVP9 libvpx EncodingMobile Neural NetworkTNNNAS Parallel BenchmarksAOM AV1BRL-CADTimed Linux Kernel CompilationsrsRANdav1dIntel Open Image DenoiseSVT-HEVC

sandy bridge 2700k 2021tnn: CPU - SqueezeNet v2ncnn: Vulkan GPU - efficientnet-b0ncnn: Vulkan GPU - mnasnetncnn: Vulkan GPU - shufflenet-v2mnn: mobilenet-v1-1.0ncnn: CPU - shufflenet-v2ncnn: Vulkan GPU - blazefacencnn: Vulkan GPU - regnety_400mnpb: EP.Cncnn: Vulkan GPU-v3-v3 - mobilenet-v3ncnn: CPU-v3-v3 - mobilenet-v3svt-av1: Preset 4 - Bosphorus 4Kncnn: CPU - efficientnet-b0ncnn: Vulkan GPU-v2-v2 - mobilenet-v2vpxenc: Speed 5 - Bosphorus 1080pncnn: CPU-v2-v2 - mobilenet-v2ncnn: Vulkan GPU - googlenettnn: CPU - DenseNetmnn: squeezenetv1.1ncnn: CPU - mnasnetncnn: CPU - resnet50srsran: 4G PHY_DL_Test 100 PRB SISO 64-QAMmnn: inception-v3srsran: 4G PHY_DL_Test 100 PRB SISO 64-QAMbuild-gdb: Time To Compilebuild-ffmpeg: Time To Compilevpxenc: Speed 5 - Bosphorus 4Kaom-av1: Speed 8 Realtime - Bosphorus 1080pncnn: CPU - squeezenet_ssdnpb: CG.Csvt-av1: Preset 4 - Bosphorus 1080pnpb: BT.Caom-av1: Speed 9 Realtime - Bosphorus 1080psrsran: 5G PHY_DL_NR Test 270 PRB SISO 256-QAMtnn: CPU - SqueezeNet v1.1svt-av1: Preset 8 - Bosphorus 1080paom-av1: Speed 9 Realtime - Bosphorus 4Ksrsran: 4G PHY_DL_Test 100 PRB MIMO 64-QAMnpb: EP.Daom-av1: Speed 6 Two-Pass - Bosphorus 4Kdav1d: Summer Nature 4Kaom-av1: Speed 8 Realtime - Bosphorus 4Kncnn: Vulkan GPU - resnet18ncnn: CPU - googlenetnpb: LU.Csrsran: 4G PHY_DL_Test 100 PRB MIMO 256-QAMaom-av1: Speed 4 Two-Pass - Bosphorus 1080pnpb: FT.Cvpxenc: Speed 0 - Bosphorus 1080ptnn: CPU - MobileNet v2srsran: 4G PHY_DL_Test 100 PRB MIMO 256-QAMncnn: CPU - regnety_400mmnn: mobilenetV3ncnn: Vulkan GPU - alexnetmnn: SqueezeNetV1.0ncnn: Vulkan GPU - vgg16ncnn: Vulkan GPU - mobilenetsrsran: 4G PHY_DL_Test 100 PRB SISO 256-QAMaom-av1: Speed 6 Two-Pass - Bosphorus 1080pdav1d: Chimera 1080p 10-bitsvt-av1: Preset 8 - Bosphorus 4Kncnn: CPU - mobilenetsrsran: 4G PHY_DL_Test 100 PRB SISO 256-QAMbrl-cad: VGR Performance Metricncnn: CPU - blazefacedav1d: Chimera 1080pncnn: CPU - vgg16ncnn: CPU - resnet18srsran: OFDM_Testaom-av1: Speed 6 Realtime - Bosphorus 4Kncnn: Vulkan GPU - yolov4-tinybuild-linux-kernel: Time To Compilesrsran: 4G PHY_DL_Test 100 PRB MIMO 64-QAMnpb: SP.Cncnn: Vulkan GPU - resnet50mnn: resnet-v2-50npb: MG.Cmnn: MobileNetV2_224npb: SP.Bdav1d: Summer Nature 1080pncnn: CPU - yolov4-tinyncnn: Vulkan GPU - squeezenet_ssdncnn: CPU - alexnetoidn: RTLightmap.hdr.4096x4096oidn: RT.ldr_alb_nrm.3840x2160oidn: RT.hdr_alb_nrm.3840x2160vpxenc: Speed 0 - Bosphorus 4Ksvt-hevc: 10 - Bosphorus 1080psvt-hevc: 7 - Bosphorus 1080psvt-hevc: 1 - Bosphorus 1080paom-av1: Speed 6 Realtime - Bosphorus 1080paom-av1: Speed 0 Two-Pass - Bosphorus 1080paom-av1: Speed 4 Two-Pass - Bosphorus 4Kaom-av1: Speed 0 Two-Pass - Bosphorus 4Ksrsran: 5G PHY_DL_NR Test 270 PRB SISO 256-QAMsrsran: 5G PHY_DL_NR Test 52 PRB SISO 64-QAMsrsran: 5G PHY_DL_NR Test 52 PRB SISO 64-QAM12393.63925.1515.129.8615.6889.873.1423.94286.3614.3114.320.04625.1215.5713.0015.5984.634922.92011.58715.15170.63261.1135.578132.3164.321175.2484.7548.76102.952223.700.1799464.6060.7211.8302.8132.40216.17257.9285.681.5048.1513.8199.0084.969322.9387.81.755509.453.83351.557287.723.954.14450.1520.767903.4061.89289.54.9753.900.57461.80149.7338163.14193.22901.7098.96678000003.43205.60293.55482.12978.35170.74111.6964404.109.9322889.92184.51205.12102.5450.180.040.070.071.495.052.450.236.080.080.790.0313.16.112.694.99524.2914.659.5615.4969.73.0623.38283.431414.150.04724.7415.2713.2515.4283.154841.40311.40515.07168.59257.3134.029130.4162.375173.044.8149.35102.022249.060.1819500.2861.2611.8304.8452.41716.14259.6284.161.548.3713.8698.3684.659268.2787.51.755479.093.85352.562287.723.924.12549.9520.824901.8861.97288.34.9853.90.57661.79149.2337063.13193902.1198.66678000003.44205.21294.01682.32979.9170.58111.5944400.319.9372889.55184.67205102.5350.180.040.070.071.495.052.450.236.080.080.790.0313.16.112.697.07924.414.829.6315.2219.613.0723.46289.7414.1714.010.04724.6315.3913.2215.383.994855.34811.4214.92171.16260.5133.61132.1162.145173.5464.7949.37101.782241.530.1819555.2161.1211.7302.3262.4216.26257.7286.251.5148.4713.998.6985.179309.3587.31.765502.83.83350.8286.323.844.1350.1820.861899.3361.71288.64.9954.090.57661.59149.6337743.13192.61899.3498.78680000003.44205.02294.34782.32972.92170.37111.5174405.919.9262892.7184.63205.17102.4950.20.040.070.071.495.052.450.236.080.080.790.0313.16.112.6OpenBenchmarking.org

TNN

TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: SqueezeNet v212320406080100SE +/- 0.27, N = 393.6495.0097.08MIN: 92.58 / MAX: 96.81MIN: 93.71 / MAX: 97.4MIN: 96.14 / MAX: 100.211. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl
OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: SqueezeNet v212320406080100Min: 93.1 / Avg: 93.64 / Max: 941. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: efficientnet-b0123612182430SE +/- 0.10, N = 325.1524.2924.40MIN: 24.6 / MAX: 64.3MIN: 23.81 / MAX: 25.07MIN: 24 / MAX: 24.921. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: efficientnet-b0123612182430Min: 25.03 / Avg: 25.15 / Max: 25.351. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: mnasnet12348121620SE +/- 0.01, N = 315.1214.6514.82MIN: 15.03 / MAX: 16.31MIN: 14.52 / MAX: 14.81MIN: 14.73 / MAX: 14.991. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: mnasnet12348121620Min: 15.11 / Avg: 15.12 / Max: 15.131. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: shufflenet-v21233691215SE +/- 0.01, N = 39.869.569.63MIN: 9.8 / MAX: 10.6MIN: 9.49 / MAX: 10.27MIN: 9.58 / MAX: 10.361. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: shufflenet-v21233691215Min: 9.85 / Avg: 9.86 / Max: 9.871. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: mobilenet-v1-1.012348121620SE +/- 0.01, N = 315.6915.5015.22MIN: 15.26 / MAX: 31.98MIN: 15 / MAX: 31.71MIN: 15.07 / MAX: 16.531. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: mobilenet-v1-1.012348121620Min: 15.67 / Avg: 15.69 / Max: 15.711. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: shufflenet-v21233691215SE +/- 0.00, N = 39.879.709.61MIN: 9.81 / MAX: 10.67MIN: 9.61 / MAX: 10.44MIN: 9.55 / MAX: 10.371. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: shufflenet-v21233691215Min: 9.87 / Avg: 9.87 / Max: 9.871. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: blazeface1230.70651.4132.11952.8263.5325SE +/- 0.00, N = 33.143.063.07MIN: 3.11 / MAX: 3.3MIN: 3.03 / MAX: 3.17MIN: 3.03 / MAX: 4.211. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: blazeface123246810Min: 3.14 / Avg: 3.14 / Max: 3.141. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: regnety_400m123612182430SE +/- 0.02, N = 323.9423.3823.46MIN: 23.83 / MAX: 37.48MIN: 23.27 / MAX: 24.52MIN: 23.39 / MAX: 24.351. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: regnety_400m123612182430Min: 23.92 / Avg: 23.94 / Max: 23.981. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.C12360120180240300SE +/- 0.59, N = 3286.36283.43289.741. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.C12350100150200250Min: 285.25 / Avg: 286.36 / Max: 287.251. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU-v3-v3 - Model: mobilenet-v312348121620SE +/- 0.01, N = 314.3114.0014.17MIN: 14.22 / MAX: 15.09MIN: 13.8 / MAX: 15.43MIN: 14.04 / MAX: 14.981. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU-v3-v3 - Model: mobilenet-v312348121620Min: 14.29 / Avg: 14.31 / Max: 14.321. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU-v3-v3 - Model: mobilenet-v312348121620SE +/- 0.00, N = 314.3214.1514.01MIN: 14.23 / MAX: 15.13MIN: 14.05 / MAX: 14.76MIN: 13.94 / MAX: 14.821. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU-v3-v3 - Model: mobilenet-v312348121620Min: 14.31 / Avg: 14.32 / Max: 14.321. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

SVT-AV1

This is a benchmark of the SVT-AV1 open-source video encoder/decoder. SVT-AV1 was originally developed by Intel as part of their Open Visual Cloud / Scalable Video Technology (SVT). Development of SVT-AV1 has since moved to the Alliance for Open Media as part of upstream AV1 development. SVT-AV1 is a CPU-based multi-threaded video encoder for the AV1 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8.7Encoder Mode: Preset 4 - Input: Bosphorus 4K1230.01060.02120.03180.04240.053SE +/- 0.000, N = 30.0460.0470.0471. (CXX) g++ options: -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8.7Encoder Mode: Preset 4 - Input: Bosphorus 4K12312345Min: 0.05 / Avg: 0.05 / Max: 0.051. (CXX) g++ options: -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: efficientnet-b0123612182430SE +/- 0.03, N = 325.1224.7424.63MIN: 24.96 / MAX: 38.31MIN: 24.53 / MAX: 25.39MIN: 24.37 / MAX: 26.551. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: efficientnet-b0123612182430Min: 25.06 / Avg: 25.12 / Max: 25.171. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU-v2-v2 - Model: mobilenet-v212348121620SE +/- 0.01, N = 315.5715.2715.39MIN: 15.48 / MAX: 16.38MIN: 15.2 / MAX: 15.44MIN: 15.29 / MAX: 16.561. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU-v2-v2 - Model: mobilenet-v212348121620Min: 15.55 / Avg: 15.57 / Max: 15.591. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

VP9 libvpx Encoding

This is a standard video encoding performance test of Google's libvpx library and the vpxenc command for the VP9 video format. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVP9 libvpx Encoding 1.10.0Speed: Speed 5 - Input: Bosphorus 1080p1233691215SE +/- 0.02, N = 313.0013.2513.221. (CXX) g++ options: -m64 -lm -lpthread -O3 -fPIC -U_FORTIFY_SOURCE -std=gnu++11
OpenBenchmarking.orgFrames Per Second, More Is BetterVP9 libvpx Encoding 1.10.0Speed: Speed 5 - Input: Bosphorus 1080p12348121620Min: 12.98 / Avg: 13 / Max: 13.031. (CXX) g++ options: -m64 -lm -lpthread -O3 -fPIC -U_FORTIFY_SOURCE -std=gnu++11

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU-v2-v2 - Model: mobilenet-v212348121620SE +/- 0.02, N = 315.5915.4215.30MIN: 15.49 / MAX: 16.81MIN: 15.23 / MAX: 15.66MIN: 15.19 / MAX: 15.651. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU-v2-v2 - Model: mobilenet-v212348121620Min: 15.56 / Avg: 15.59 / Max: 15.631. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: googlenet12320406080100SE +/- 0.09, N = 384.6383.1583.99MIN: 83.08 / MAX: 93.81MIN: 81.77 / MAX: 84.23MIN: 81.96 / MAX: 96.861. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: googlenet1231632486480Min: 84.48 / Avg: 84.63 / Max: 84.791. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

TNN

TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: DenseNet12311002200330044005500SE +/- 7.66, N = 34922.924841.404855.35MIN: 4889.71 / MAX: 4977.94MIN: 4821.23 / MAX: 4862.58MIN: 4841.57 / MAX: 4875.251. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl
OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: DenseNet1239001800270036004500Min: 4908.97 / Avg: 4922.92 / Max: 4935.381. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: squeezenetv1.11233691215SE +/- 0.05, N = 311.5911.4111.42MIN: 11.41 / MAX: 29.09MIN: 11.1 / MAX: 28.09MIN: 11.36 / MAX: 12.131. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: squeezenetv1.11233691215Min: 11.49 / Avg: 11.59 / Max: 11.651. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: mnasnet12348121620SE +/- 0.01, N = 315.1515.0714.92MIN: 15.06 / MAX: 15.76MIN: 14.85 / MAX: 15.3MIN: 14.75 / MAX: 15.151. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: mnasnet12348121620Min: 15.14 / Avg: 15.15 / Max: 15.161. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: resnet501234080120160200SE +/- 0.11, N = 3170.63168.59171.16MIN: 169.33 / MAX: 182.41MIN: 166.07 / MAX: 179.43MIN: 170.11 / MAX: 172.531. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: resnet50123306090120150Min: 170.41 / Avg: 170.63 / Max: 170.761. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB SISO 64-QAM12360120180240300SE +/- 0.26, N = 3261.1257.3260.51. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto
OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB SISO 64-QAM12350100150200250Min: 260.6 / Avg: 261.07 / Max: 261.51. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: inception-v3123306090120150SE +/- 0.70, N = 3135.58134.03133.61MIN: 133.31 / MAX: 181.95MIN: 133.3 / MAX: 149.57MIN: 133.26 / MAX: 149.971. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: inception-v3123306090120150Min: 134.17 / Avg: 135.58 / Max: 136.341. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB SISO 64-QAM123306090120150SE +/- 0.07, N = 3132.3130.4132.11. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto
OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB SISO 64-QAM12320406080100Min: 132.2 / Avg: 132.27 / Max: 132.41. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto

Timed GDB GNU Debugger Compilation

This test times how long it takes to build the GNU Debugger (GDB) in a default configuration. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed GDB GNU Debugger Compilation 10.2Time To Compile1234080120160200SE +/- 0.13, N = 3164.32162.38162.15
OpenBenchmarking.orgSeconds, Fewer Is BetterTimed GDB GNU Debugger Compilation 10.2Time To Compile123306090120150Min: 164.11 / Avg: 164.32 / Max: 164.55

Timed FFmpeg Compilation

This test times how long it takes to build the FFmpeg multimedia library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed FFmpeg Compilation 4.4Time To Compile1234080120160200SE +/- 0.38, N = 3175.25173.04173.55
OpenBenchmarking.orgSeconds, Fewer Is BetterTimed FFmpeg Compilation 4.4Time To Compile123306090120150Min: 174.54 / Avg: 175.25 / Max: 175.83

VP9 libvpx Encoding

This is a standard video encoding performance test of Google's libvpx library and the vpxenc command for the VP9 video format. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVP9 libvpx Encoding 1.10.0Speed: Speed 5 - Input: Bosphorus 4K1231.08232.16463.24694.32925.4115SE +/- 0.03, N = 34.754.814.791. (CXX) g++ options: -m64 -lm -lpthread -O3 -fPIC -U_FORTIFY_SOURCE -std=gnu++11
OpenBenchmarking.orgFrames Per Second, More Is BetterVP9 libvpx Encoding 1.10.0Speed: Speed 5 - Input: Bosphorus 4K123246810Min: 4.7 / Avg: 4.75 / Max: 4.781. (CXX) g++ options: -m64 -lm -lpthread -O3 -fPIC -U_FORTIFY_SOURCE -std=gnu++11

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 8 Realtime - Input: Bosphorus 1080p1231122334455SE +/- 0.16, N = 348.7649.3549.371. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 8 Realtime - Input: Bosphorus 1080p1231020304050Min: 48.44 / Avg: 48.76 / Max: 48.931. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: squeezenet_ssd12320406080100SE +/- 0.07, N = 3102.95102.02101.78MIN: 102.25 / MAX: 113.35MIN: 101.35 / MAX: 104.02MIN: 101.22 / MAX: 108.551. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: squeezenet_ssd12320406080100Min: 102.85 / Avg: 102.95 / Max: 103.081. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: CG.C1235001000150020002500SE +/- 2.59, N = 32223.702249.062241.531. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: CG.C123400800120016002000Min: 2219.42 / Avg: 2223.7 / Max: 2228.381. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0

SVT-AV1

This is a benchmark of the SVT-AV1 open-source video encoder/decoder. SVT-AV1 was originally developed by Intel as part of their Open Visual Cloud / Scalable Video Technology (SVT). Development of SVT-AV1 has since moved to the Alliance for Open Media as part of upstream AV1 development. SVT-AV1 is a CPU-based multi-threaded video encoder for the AV1 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8.7Encoder Mode: Preset 4 - Input: Bosphorus 1080p1230.04070.08140.12210.16280.2035SE +/- 0.000, N = 30.1790.1810.1811. (CXX) g++ options: -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8.7Encoder Mode: Preset 4 - Input: Bosphorus 1080p12312345Min: 0.18 / Avg: 0.18 / Max: 0.181. (CXX) g++ options: -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: BT.C1232K4K6K8K10KSE +/- 13.47, N = 39464.609500.289555.211. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: BT.C12317003400510068008500Min: 9448.8 / Avg: 9464.6 / Max: 9491.391. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 9 Realtime - Input: Bosphorus 1080p1231428425670SE +/- 0.02, N = 360.7261.2661.121. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 9 Realtime - Input: Bosphorus 1080p1231224364860Min: 60.69 / Avg: 60.72 / Max: 60.771. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 21.04Test: 5G PHY_DL_NR Test 270 PRB SISO 256-QAM1233691215SE +/- 0.00, N = 311.811.811.71. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto
OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 21.04Test: 5G PHY_DL_NR Test 270 PRB SISO 256-QAM1233691215Min: 11.8 / Avg: 11.8 / Max: 11.81. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto

TNN

TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: SqueezeNet v1.112370140210280350SE +/- 1.24, N = 3302.81304.85302.33MIN: 296.54 / MAX: 333.79MIN: 301.55 / MAX: 308.36MIN: 300.89 / MAX: 304.681. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl
OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: SqueezeNet v1.112350100150200250Min: 300.7 / Avg: 302.81 / Max: 304.991. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl

SVT-AV1

This is a benchmark of the SVT-AV1 open-source video encoder/decoder. SVT-AV1 was originally developed by Intel as part of their Open Visual Cloud / Scalable Video Technology (SVT). Development of SVT-AV1 has since moved to the Alliance for Open Media as part of upstream AV1 development. SVT-AV1 is a CPU-based multi-threaded video encoder for the AV1 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8.7Encoder Mode: Preset 8 - Input: Bosphorus 1080p1230.54451.0891.63352.1782.7225SE +/- 0.001, N = 32.4022.4172.4201. (CXX) g++ options: -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8.7Encoder Mode: Preset 8 - Input: Bosphorus 1080p123246810Min: 2.4 / Avg: 2.4 / Max: 2.41. (CXX) g++ options: -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 9 Realtime - Input: Bosphorus 4K12348121620SE +/- 0.01, N = 316.1716.1416.261. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 9 Realtime - Input: Bosphorus 4K12348121620Min: 16.15 / Avg: 16.17 / Max: 16.191. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB MIMO 64-QAM12360120180240300SE +/- 0.62, N = 3257.9259.6257.71. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto
OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB MIMO 64-QAM12350100150200250Min: 256.7 / Avg: 257.93 / Max: 258.71. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.D12360120180240300SE +/- 0.34, N = 3285.68284.16286.251. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.D12350100150200250Min: 285.11 / Avg: 285.68 / Max: 286.291. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 4K1230.33980.67961.01941.35921.699SE +/- 0.00, N = 31.501.501.511. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 4K123246810Min: 1.49 / Avg: 1.5 / Max: 1.51. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.9.0Video Input: Summer Nature 4K1231122334455SE +/- 0.15, N = 348.1548.3748.47MIN: 45.52 / MAX: 52.41MIN: 45.9 / MAX: 52.25MIN: 46.06 / MAX: 52.251. (CC) gcc options: -pthread -lm
OpenBenchmarking.orgFPS, More Is Betterdav1d 0.9.0Video Input: Summer Nature 4K1231020304050Min: 47.97 / Avg: 48.15 / Max: 48.451. (CC) gcc options: -pthread -lm

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 8 Realtime - Input: Bosphorus 4K12348121620SE +/- 0.05, N = 313.8113.8613.901. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 8 Realtime - Input: Bosphorus 4K12348121620Min: 13.72 / Avg: 13.81 / Max: 13.881. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: resnet1812320406080100SE +/- 0.06, N = 399.0098.3698.69MIN: 98.01 / MAX: 110.34MIN: 97.14 / MAX: 99.75MIN: 97.96 / MAX: 100.241. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: resnet1812320406080100Min: 98.9 / Avg: 99 / Max: 99.11. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: googlenet12320406080100SE +/- 0.08, N = 384.9684.6585.17MIN: 83.92 / MAX: 151.07MIN: 83.33 / MAX: 86.4MIN: 84.2 / MAX: 88.591. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: googlenet1231632486480Min: 84.88 / Avg: 84.96 / Max: 85.111. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: LU.C1232K4K6K8K10KSE +/- 22.62, N = 39322.939268.279309.351. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: LU.C12316003200480064008000Min: 9298.19 / Avg: 9322.93 / Max: 9368.11. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB MIMO 256-QAM12320406080100SE +/- 0.18, N = 387.887.587.31. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto
OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB MIMO 256-QAM12320406080100Min: 87.5 / Avg: 87.83 / Max: 88.11. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 1080p1230.3960.7921.1881.5841.98SE +/- 0.00, N = 31.751.751.761. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 1080p123246810Min: 1.74 / Avg: 1.75 / Max: 1.751. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: FT.C12312002400360048006000SE +/- 20.22, N = 35509.455479.095502.801. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: FT.C12310002000300040005000Min: 5474.13 / Avg: 5509.45 / Max: 5544.151. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0

VP9 libvpx Encoding

This is a standard video encoding performance test of Google's libvpx library and the vpxenc command for the VP9 video format. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVP9 libvpx Encoding 1.10.0Speed: Speed 0 - Input: Bosphorus 1080p1230.86631.73262.59893.46524.3315SE +/- 0.01, N = 33.833.853.831. (CXX) g++ options: -m64 -lm -lpthread -O3 -fPIC -U_FORTIFY_SOURCE -std=gnu++11
OpenBenchmarking.orgFrames Per Second, More Is BetterVP9 libvpx Encoding 1.10.0Speed: Speed 0 - Input: Bosphorus 1080p123246810Min: 3.81 / Avg: 3.83 / Max: 3.851. (CXX) g++ options: -m64 -lm -lpthread -O3 -fPIC -U_FORTIFY_SOURCE -std=gnu++11

TNN

TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: MobileNet v212380160240320400SE +/- 0.20, N = 3351.56352.56350.80MIN: 348.02 / MAX: 359MIN: 350.42 / MAX: 360.89MIN: 349.48 / MAX: 354.051. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl
OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: MobileNet v212360120180240300Min: 351.16 / Avg: 351.56 / Max: 351.761. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB MIMO 256-QAM12360120180240300SE +/- 0.18, N = 3287.7287.7286.31. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto
OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB MIMO 256-QAM12350100150200250Min: 287.4 / Avg: 287.67 / Max: 2881. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: regnety_400m123612182430SE +/- 0.03, N = 323.9523.9223.84MIN: 23.83 / MAX: 38.28MIN: 23.61 / MAX: 24.4MIN: 23.65 / MAX: 24.731. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: regnety_400m123612182430Min: 23.91 / Avg: 23.95 / Max: 24.011. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: mobilenetV31230.93241.86482.79723.72964.662SE +/- 0.004, N = 34.1444.1254.130MIN: 4.12 / MAX: 6.51MIN: 4.08 / MAX: 6.35MIN: 4.1 / MAX: 5.151. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: mobilenetV3123246810Min: 4.14 / Avg: 4.14 / Max: 4.151. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: alexnet1231122334455SE +/- 0.06, N = 350.1549.9550.18MIN: 49.48 / MAX: 52.9MIN: 49.38 / MAX: 62.12MIN: 49.85 / MAX: 50.931. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: alexnet1231020304050Min: 50.04 / Avg: 50.15 / Max: 50.251. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: SqueezeNetV1.0123510152025SE +/- 0.04, N = 320.7720.8220.86MIN: 20.59 / MAX: 37.58MIN: 20.73 / MAX: 36.9MIN: 20.79 / MAX: 21.671. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: SqueezeNetV1.0123510152025Min: 20.68 / Avg: 20.77 / Max: 20.831. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: vgg161232004006008001000SE +/- 2.16, N = 3903.40901.88899.33MIN: 885.02 / MAX: 954.01MIN: 890.6 / MAX: 918.96MIN: 889.42 / MAX: 910.381. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: vgg16123160320480640800Min: 899.68 / Avg: 903.4 / Max: 907.151. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: mobilenet1231428425670SE +/- 0.13, N = 361.8961.9761.71MIN: 61.46 / MAX: 110.53MIN: 61.67 / MAX: 75.55MIN: 61.54 / MAX: 62.531. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: mobilenet1231224364860Min: 61.69 / Avg: 61.89 / Max: 62.131. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB SISO 256-QAM12360120180240300SE +/- 0.48, N = 3289.5288.3288.61. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto
OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB SISO 256-QAM12350100150200250Min: 288.6 / Avg: 289.53 / Max: 290.21. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 1080p1231.12282.24563.36844.49125.614SE +/- 0.01, N = 34.974.984.991. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 1080p123246810Min: 4.96 / Avg: 4.97 / Max: 4.981. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.9.0Video Input: Chimera 1080p 10-bit1231224364860SE +/- 0.12, N = 353.9053.9054.09MIN: 36.53 / MAX: 121.89MIN: 36.55 / MAX: 121.55MIN: 36.69 / MAX: 121.131. (CC) gcc options: -pthread -lm
OpenBenchmarking.orgFPS, More Is Betterdav1d 0.9.0Video Input: Chimera 1080p 10-bit1231122334455Min: 53.66 / Avg: 53.9 / Max: 54.041. (CC) gcc options: -pthread -lm

SVT-AV1

This is a benchmark of the SVT-AV1 open-source video encoder/decoder. SVT-AV1 was originally developed by Intel as part of their Open Visual Cloud / Scalable Video Technology (SVT). Development of SVT-AV1 has since moved to the Alliance for Open Media as part of upstream AV1 development. SVT-AV1 is a CPU-based multi-threaded video encoder for the AV1 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8.7Encoder Mode: Preset 8 - Input: Bosphorus 4K1230.12960.25920.38880.51840.648SE +/- 0.000, N = 30.5740.5760.5761. (CXX) g++ options: -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.8.7Encoder Mode: Preset 8 - Input: Bosphorus 4K123246810Min: 0.57 / Avg: 0.57 / Max: 0.581. (CXX) g++ options: -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: mobilenet1231428425670SE +/- 0.02, N = 361.8061.7961.59MIN: 61.61 / MAX: 63.46MIN: 61.62 / MAX: 63.24MIN: 61.38 / MAX: 62.321. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: mobilenet1231224364860Min: 61.77 / Avg: 61.8 / Max: 61.831. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB SISO 256-QAM123306090120150SE +/- 0.03, N = 3149.7149.2149.61. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto
OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB SISO 256-QAM123306090120150Min: 149.7 / Avg: 149.73 / Max: 149.81. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto

BRL-CAD

BRL-CAD is a cross-platform, open-source solid modeling system with built-in benchmark mode. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgVGR Performance Metric, More Is BetterBRL-CAD 7.32.2VGR Performance Metric1237K14K21K28K35K3381633706337741. (CXX) g++ options: -std=c++11 -pipe -fvisibility=hidden -fno-strict-aliasing -fno-common -fexceptions -ftemplate-depth-128 -m64 -ggdb3 -O3 -fipa-pta -fstrength-reduce -finline-functions -flto -pedantic -pthread -ldl -lm

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: blazeface1230.70651.4132.11952.8263.5325SE +/- 0.00, N = 33.143.133.13MIN: 3.1 / MAX: 4.01MIN: 3.1 / MAX: 3.28MIN: 3.04 / MAX: 3.311. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: blazeface123246810Min: 3.14 / Avg: 3.14 / Max: 3.141. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.9.0Video Input: Chimera 1080p1234080120160200SE +/- 0.13, N = 3193.22193.00192.61MIN: 143.92 / MAX: 339.04MIN: 144.66 / MAX: 336.59MIN: 144.86 / MAX: 339.721. (CC) gcc options: -pthread -lm
OpenBenchmarking.orgFPS, More Is Betterdav1d 0.9.0Video Input: Chimera 1080p1234080120160200Min: 193.02 / Avg: 193.22 / Max: 193.461. (CC) gcc options: -pthread -lm

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: vgg161232004006008001000SE +/- 0.54, N = 3901.70902.11899.34MIN: 885.66 / MAX: 1147.12MIN: 891.03 / MAX: 914.02MIN: 888.07 / MAX: 913.461. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: vgg16123160320480640800Min: 900.74 / Avg: 901.7 / Max: 902.61. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: resnet1812320406080100SE +/- 0.03, N = 398.9698.6698.78MIN: 98.1 / MAX: 100.89MIN: 97.13 / MAX: 110.39MIN: 97.76 / MAX: 100.361. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: resnet1812320406080100Min: 98.9 / Avg: 98.96 / Max: 99.021. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSamples / Second, More Is BettersrsRAN 21.04Test: OFDM_Test12315M30M45M60M75MSE +/- 461880.22, N = 36780000067800000680000001. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto
OpenBenchmarking.orgSamples / Second, More Is BettersrsRAN 21.04Test: OFDM_Test12312M24M36M48M60MMin: 67000000 / Avg: 67800000 / Max: 686000001. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 6 Realtime - Input: Bosphorus 4K1230.7741.5482.3223.0963.87SE +/- 0.00, N = 33.433.443.441. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 6 Realtime - Input: Bosphorus 4K123246810Min: 3.43 / Avg: 3.43 / Max: 3.431. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: yolov4-tiny12350100150200250SE +/- 0.20, N = 3205.60205.21205.02MIN: 203.56 / MAX: 215.57MIN: 203.4 / MAX: 209.04MIN: 203.49 / MAX: 213.981. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: yolov4-tiny1234080120160200Min: 205.34 / Avg: 205.6 / Max: 205.981. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

Timed Linux Kernel Compilation

This test times how long it takes to build the Linux kernel in a default configuration (defconfig) for the architecture being tested. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 5.10.20Time To Compile12360120180240300SE +/- 0.83, N = 3293.55294.02294.35
OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 5.10.20Time To Compile12350100150200250Min: 292.67 / Avg: 293.55 / Max: 295.21

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB MIMO 64-QAM12320406080100SE +/- 0.09, N = 382.182.382.31. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto
OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 21.04Test: 4G PHY_DL_Test 100 PRB MIMO 64-QAM1231632486480Min: 82 / Avg: 82.13 / Max: 82.31. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: SP.C1236001200180024003000SE +/- 2.80, N = 32978.352979.902972.921. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: SP.C1235001000150020002500Min: 2973.41 / Avg: 2978.35 / Max: 2983.091. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: resnet501234080120160200SE +/- 0.27, N = 3170.74170.58170.37MIN: 166.72 / MAX: 183.83MIN: 166.65 / MAX: 172.58MIN: 166.83 / MAX: 178.911. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: resnet50123306090120150Min: 170.21 / Avg: 170.74 / Max: 171.061. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: resnet-v2-50123306090120150SE +/- 0.33, N = 3111.70111.59111.52MIN: 110.77 / MAX: 144.54MIN: 108.18 / MAX: 140.38MIN: 108.2 / MAX: 151.021. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: resnet-v2-5012320406080100Min: 111.24 / Avg: 111.7 / Max: 112.341. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: MG.C1239001800270036004500SE +/- 0.77, N = 34404.104400.314405.911. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: MG.C1238001600240032004000Min: 4402.56 / Avg: 4404.1 / Max: 4404.951. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: MobileNetV2_2241233691215SE +/- 0.009, N = 39.9329.9379.926MIN: 9.88 / MAX: 12.28MIN: 9.87 / MAX: 15.49MIN: 9.86 / MAX: 11.51. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: MobileNetV2_2241233691215Min: 9.92 / Avg: 9.93 / Max: 9.941. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: SP.B1236001200180024003000SE +/- 0.11, N = 32889.922889.552892.701. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0
OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: SP.B1235001000150020002500Min: 2889.71 / Avg: 2889.92 / Max: 2890.041. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -ldl -levent_core -levent_pthreads -lutil -lm -lrt -lz 2. Open MPI 4.1.0

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.9.0Video Input: Summer Nature 1080p1234080120160200SE +/- 0.21, N = 3184.51184.67184.63MIN: 164.08 / MAX: 202.12MIN: 164.54 / MAX: 201.27MIN: 164.89 / MAX: 201.041. (CC) gcc options: -pthread -lm
OpenBenchmarking.orgFPS, More Is Betterdav1d 0.9.0Video Input: Summer Nature 1080p123306090120150Min: 184.15 / Avg: 184.51 / Max: 184.891. (CC) gcc options: -pthread -lm

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: yolov4-tiny12350100150200250SE +/- 0.08, N = 3205.12205.00205.17MIN: 203.01 / MAX: 225.06MIN: 203.46 / MAX: 207.58MIN: 203.76 / MAX: 211.881. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: yolov4-tiny1234080120160200Min: 204.97 / Avg: 205.12 / Max: 205.221. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: squeezenet_ssd12320406080100SE +/- 0.24, N = 3102.54102.53102.49MIN: 101.11 / MAX: 106.09MIN: 101.19 / MAX: 113.04MIN: 101.39 / MAX: 105.511. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: Vulkan GPU - Model: squeezenet_ssd12320406080100Min: 102.06 / Avg: 102.54 / Max: 102.841. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: alexnet1231122334455SE +/- 0.06, N = 350.1850.1850.20MIN: 49.74 / MAX: 59.36MIN: 49.51 / MAX: 51.39MIN: 49.7 / MAX: 51.751. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210525Target: CPU - Model: alexnet1231020304050Min: 50.07 / Avg: 50.18 / Max: 50.291. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

Intel Open Image Denoise

Open Image Denoise is a denoising library for ray-tracing and part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 1.4.0Run: RTLightmap.hdr.4096x40961230.0090.0180.0270.0360.045SE +/- 0.00, N = 30.040.040.04
OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 1.4.0Run: RTLightmap.hdr.4096x409612312345Min: 0.04 / Avg: 0.04 / Max: 0.04

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 1.4.0Run: RT.ldr_alb_nrm.3840x21601230.01580.03160.04740.06320.079SE +/- 0.00, N = 30.070.070.07
OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 1.4.0Run: RT.ldr_alb_nrm.3840x216012312345Min: 0.07 / Avg: 0.07 / Max: 0.07

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 1.4.0Run: RT.hdr_alb_nrm.3840x21601230.01580.03160.04740.06320.079SE +/- 0.00, N = 30.070.070.07
OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 1.4.0Run: RT.hdr_alb_nrm.3840x216012312345Min: 0.07 / Avg: 0.07 / Max: 0.07

VP9 libvpx Encoding

This is a standard video encoding performance test of Google's libvpx library and the vpxenc command for the VP9 video format. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVP9 libvpx Encoding 1.10.0Speed: Speed 0 - Input: Bosphorus 4K1230.33530.67061.00591.34121.6765SE +/- 0.00, N = 31.491.491.491. (CXX) g++ options: -m64 -lm -lpthread -O3 -fPIC -U_FORTIFY_SOURCE -std=gnu++11
OpenBenchmarking.orgFrames Per Second, More Is BetterVP9 libvpx Encoding 1.10.0Speed: Speed 0 - Input: Bosphorus 4K123246810Min: 1.48 / Avg: 1.49 / Max: 1.491. (CXX) g++ options: -m64 -lm -lpthread -O3 -fPIC -U_FORTIFY_SOURCE -std=gnu++11

SVT-HEVC

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 10 - Input: Bosphorus 1080p1231.13632.27263.40894.54525.6815SE +/- 0.00, N = 35.055.055.051. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 10 - Input: Bosphorus 1080p123246810Min: 5.05 / Avg: 5.05 / Max: 5.051. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 7 - Input: Bosphorus 1080p1230.55131.10261.65392.20522.7565SE +/- 0.00, N = 32.452.452.451. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 7 - Input: Bosphorus 1080p123246810Min: 2.45 / Avg: 2.45 / Max: 2.451. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 1 - Input: Bosphorus 1080p1230.05180.10360.15540.20720.259SE +/- 0.00, N = 30.230.230.231. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 1 - Input: Bosphorus 1080p12312345Min: 0.23 / Avg: 0.23 / Max: 0.231. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 6 Realtime - Input: Bosphorus 1080p123246810SE +/- 0.01, N = 36.086.086.081. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 6 Realtime - Input: Bosphorus 1080p123246810Min: 6.06 / Avg: 6.08 / Max: 6.11. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 1080p1230.0180.0360.0540.0720.09SE +/- 0.00, N = 30.080.080.081. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 1080p12312345Min: 0.08 / Avg: 0.08 / Max: 0.081. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 4K1230.17780.35560.53340.71120.889SE +/- 0.00, N = 30.790.790.791. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 4K123246810Min: 0.78 / Avg: 0.79 / Max: 0.791. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 4K1230.00680.01360.02040.02720.034SE +/- 0.00, N = 30.030.030.031. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.1Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 4K12312345Min: 0.03 / Avg: 0.03 / Max: 0.031. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 21.04Test: 5G PHY_DL_NR Test 270 PRB SISO 256-QAM1233691215SE +/- 0.03, N = 313.113.113.11. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto
OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 21.04Test: 5G PHY_DL_NR Test 270 PRB SISO 256-QAM12348121620Min: 13.1 / Avg: 13.13 / Max: 13.21. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 21.04Test: 5G PHY_DL_NR Test 52 PRB SISO 64-QAM123246810SE +/- 0.00, N = 36.16.16.11. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto
OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 21.04Test: 5G PHY_DL_NR Test 52 PRB SISO 64-QAM123246810Min: 6.1 / Avg: 6.1 / Max: 6.11. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 21.04Test: 5G PHY_DL_NR Test 52 PRB SISO 64-QAM1233691215SE +/- 0.00, N = 312.612.612.61. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto
OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 21.04Test: 5G PHY_DL_NR Test 52 PRB SISO 64-QAM12348121620Min: 12.6 / Avg: 12.6 / Max: 12.61. (CXX) g++ options: -std=c++11 -fno-strict-aliasing -march=native -mfpmath=sse -mavx -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -rdynamic -lpthread -lbladeRF -lm -lfftw3f -lmbedcrypto

97 Results Shown

TNN
NCNN:
  Vulkan GPU - efficientnet-b0
  Vulkan GPU - mnasnet
  Vulkan GPU - shufflenet-v2
Mobile Neural Network
NCNN:
  CPU - shufflenet-v2
  Vulkan GPU - blazeface
  Vulkan GPU - regnety_400m
NAS Parallel Benchmarks
NCNN:
  Vulkan GPU-v3-v3 - mobilenet-v3
  CPU-v3-v3 - mobilenet-v3
SVT-AV1
NCNN:
  CPU - efficientnet-b0
  Vulkan GPU-v2-v2 - mobilenet-v2
VP9 libvpx Encoding
NCNN:
  CPU-v2-v2 - mobilenet-v2
  Vulkan GPU - googlenet
TNN
Mobile Neural Network
NCNN:
  CPU - mnasnet
  CPU - resnet50
srsRAN
Mobile Neural Network
srsRAN
Timed GDB GNU Debugger Compilation
Timed FFmpeg Compilation
VP9 libvpx Encoding
AOM AV1
NCNN
NAS Parallel Benchmarks
SVT-AV1
NAS Parallel Benchmarks
AOM AV1
srsRAN
TNN
SVT-AV1
AOM AV1
srsRAN
NAS Parallel Benchmarks
AOM AV1
dav1d
AOM AV1
NCNN:
  Vulkan GPU - resnet18
  CPU - googlenet
NAS Parallel Benchmarks
srsRAN
AOM AV1
NAS Parallel Benchmarks
VP9 libvpx Encoding
TNN
srsRAN
NCNN
Mobile Neural Network
NCNN
Mobile Neural Network
NCNN:
  Vulkan GPU - vgg16
  Vulkan GPU - mobilenet
srsRAN
AOM AV1
dav1d
SVT-AV1
NCNN
srsRAN
BRL-CAD
NCNN
dav1d
NCNN:
  CPU - vgg16
  CPU - resnet18
srsRAN
AOM AV1
NCNN
Timed Linux Kernel Compilation
srsRAN
NAS Parallel Benchmarks
NCNN
Mobile Neural Network
NAS Parallel Benchmarks
Mobile Neural Network
NAS Parallel Benchmarks
dav1d
NCNN:
  CPU - yolov4-tiny
  Vulkan GPU - squeezenet_ssd
  CPU - alexnet
Intel Open Image Denoise:
  RTLightmap.hdr.4096x4096
  RT.ldr_alb_nrm.3840x2160
  RT.hdr_alb_nrm.3840x2160
VP9 libvpx Encoding
SVT-HEVC:
  10 - Bosphorus 1080p
  7 - Bosphorus 1080p
  1 - Bosphorus 1080p
AOM AV1:
  Speed 6 Realtime - Bosphorus 1080p
  Speed 0 Two-Pass - Bosphorus 1080p
  Speed 4 Two-Pass - Bosphorus 4K
  Speed 0 Two-Pass - Bosphorus 4K
srsRAN:
  5G PHY_DL_NR Test 270 PRB SISO 256-QAM
  5G PHY_DL_NR Test 52 PRB SISO 64-QAM
  5G PHY_DL_NR Test 52 PRB SISO 64-QAM