AMD Ryzen 9 7950X AVX-512

AMD Ryzen 9 7950X AVX-512 benchmark comparison by Michael Larabel launch day embargo lift review. Stock/out-of-the-box build with AVX-512. For lack of any AVX-512 toggle from the ASUS BIOS, the AVX2 / non-AVX-512 run was carried out by booting kernel with "clearcpuid=304" to clear AVX-512 support from the kernel and for the binary programs that scan /proc/cpuinfo for avx512* extensions. Plus for the open-source benchmarks specifying CFLAGS/CXXFLAGS without AVX-512 extensions. See full launch day review @ https://www.phoronix.com/review/amd-zen4-avx512

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2209253-NE-RYZEN795065
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results
Show Result Confidence Charts
Allow Limiting Results To Certain Suite(s)

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs
No Box Plots
On Line Graphs With Missing Data, Connect The Line Gaps

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs
Condense Test Profiles With Multiple Version Results Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Toggle/Hide
Result
Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
Default, AVX-512 Enabled
September 17 2022
  4 Hours, 11 Minutes
Without AVX-512
September 17 2022
  3 Hours, 57 Minutes
Invert Behavior (Only Show Selected Data)
  4 Hours, 4 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


AMD Ryzen 9 7950X AVX-512OpenBenchmarking.orgPhoronix Test SuiteAMD Ryzen 9 7950X 16-Core @ 5.88GHz (16 Cores / 32 Threads)ASUS ROG CROSSHAIR X670E HERO (0604 BIOS)AMD Device 14d832GB2000GB Samsung SSD 980 PRO 2TB + 2000GBAMD Radeon RX 6800 XT 16GB (2575/1000MHz)AMD Navi 21 HDMI AudioASUS VP28UIntel I225-V + Intel Wi-Fi 6 AX210/AX211/AX411Ubuntu 22.046.0.0-060000rc1daily20220820-generic (x86_64)GNOME Shell 42.2X Server + Wayland4.6 Mesa 22.3.0-devel (git-4685385 2022-08-23 jammy-oibaf-ppa) (LLVM 14.0.6 DRM 3.48)1.3.224GCC 12.0.1 20220319ext43840x2160ProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerOpenGLVulkanCompilerFile-SystemScreen ResolutionAMD Ryzen 9 7950X AVX-512 BenchmarksSystem Logs- Transparent Huge Pages: madvise- Default, AVX-512 Enabled: CXXFLAGS="-O3 -march=native -mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi -mprefer-vector-width=512" CFLAGS="-O3 -march=native -mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi -mprefer-vector-width=512" - Without AVX-512: CXXFLAGS="-O3 -march=native -mno-avx512f" CFLAGS="-O3 -march=native -mno-avx512f"- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-12-OcsLtf/gcc-12-12-20220319/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-12-OcsLtf/gcc-12-12-20220319/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: amd-pstate performance (Boost: Enabled) - CPU Microcode: 0xa601203- Python 3.10.4- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected

Default, AVX-512 Enabled vs. Without AVX-512 ComparisonPhoronix Test SuiteBaseline+66.4%+66.4%+132.8%+132.8%+199.2%+199.2%+265.6%+265.6%Myriad-Groestl265.7%W.P.D.F.I - CPU117.1%W.P.D.F.I - CPU117%F.D.F - CPU115.1%F.D.F - CPU114.4%W.P.D.F - CPU113.7%W.P.D.F - CPU113.6%F.D.F.I - CPU109.1%F.D.F.I - CPU108.6%Blake-2 S104.5%M.T.E.T.D.F - CPU103.8%M.T.E.T.D.F - CPU103.8%A.G.R.R.0.F - CPU100%A.G.R.R.0.F - CPU100%V.D.F.I - CPU98.2%V.D.F.I - CPU98.1%LBC, LBRY Credits97.1%D.B.s - u8s8f32 - CPU92.4%D.B.s - u8s8f32 - CPU88%P.V.B.D.F - CPU87%P.V.B.D.F - CPU86.8%gravity_spheres_volume/dim_512/scivis/real_time83.1%P.D.F - CPU80.2%P.D.F - CPU79.9%P.D.F - CPU79.8%P.D.F - CPU78.9%gravity_spheres_volume/dim_512/ao/real_time77.6%A.G.R.R.0.F.I - CPU70.8%A.G.R.R.0.F.I - CPU68.2%Q.S.2.P66.9%CPU - vision_transformer58.6%R.N.N.I - u8s8f32 - CPU56.1%R.N.N.I - bf16bf16bf16 - CPU54.8%Garlicoin53.1%IP Shapes 1D - u8s8f32 - CPU50%V.D.F - CPU48%V.D.F - CPU47.9%C.B.S.A - u8s8f32 - CPU40%Skeincoin39.3%R.N.N.T - u8s8f32 - CPU37.7%gravity_spheres_volume/dim_512/pathtracer/real_time37.4%R.N.N.T - bf16bf16bf16 - CPU35.7%DistinctUserID32.6%vklBenchmark ISPC30.1%TopTweet26.1%PartialTweets25.4%1 - 4K - 1 - Path Tracer23%3 - 4K - 1 - Path Tracer22.4%Kostya21.4%1 - 4K - 32 - Path Tracer21.1%Pathtracer ISPC - Crown21%1 - 4K - 16 - Path Tracer20.7%3 - 4K - 32 - Path Tracer20.1%3 - 4K - 16 - Path Tracer20%Pathtracer ISPC - Asian Dragon15.6%LargeRand14%SqueezeNetV1.012.8%Eigen9.9%particle_volume/pathtracer/real_time8.3%Summer Nature 4K3.4%S.N.12.8%Cpuminer-OptOpenVINOOpenVINOOpenVINOOpenVINOOpenVINOOpenVINOOpenVINOOpenVINOCpuminer-OptOpenVINOOpenVINOOpenVINOOpenVINOOpenVINOOpenVINOCpuminer-OptoneDNNoneDNNOpenVINOOpenVINOOSPRayOpenVINOOpenVINOOpenVINOOpenVINOOSPRayOpenVINOOpenVINOCpuminer-OptNCNNoneDNNoneDNNCpuminer-OptoneDNNOpenVINOOpenVINOoneDNNCpuminer-OptoneDNNOSPRayoneDNNsimdjsonOpenVKLsimdjsonsimdjsonOSPRay StudioOSPRay StudiosimdjsonOSPRay StudioEmbreeOSPRay StudioOSPRay StudioOSPRay StudioEmbreesimdjsonMobile Neural NetworkLeelaChessZeroOSPRaydav1ddav1dDefault, AVX-512 EnabledWithout AVX-512

AMD Ryzen 9 7950X AVX-512dav1d: Summer Nature 4Kdav1d: Summer Nature 1080popenvino: Face Detection FP16 - CPUopenvino: Person Detection FP16 - CPUopenvino: Person Detection FP32 - CPUopenvino: Vehicle Detection FP16 - CPUopenvino: Face Detection FP16-INT8 - CPUopenvino: Vehicle Detection FP16-INT8 - CPUopenvino: Weld Porosity Detection FP16 - CPUopenvino: Machine Translation EN To DE FP16 - CPUopenvino: Weld Porosity Detection FP16-INT8 - CPUopenvino: Person Vehicle Bike Detection FP16 - CPUopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUopenvino: Age Gender Recognition Retail 0013 FP16 - CPUembree: Pathtracer ISPC - Crownembree: Pathtracer ISPC - Asian Dragonsimdjson: TopTweetsimdjson: LargeRandsimdjson: PartialTweetssimdjson: DistinctUserIDsimdjson: Kostyaopenvkl: vklBenchmark ISPCospray: particle_volume/pathtracer/real_timeospray: gravity_spheres_volume/dim_512/ao/real_timeospray: gravity_spheres_volume/dim_512/scivis/real_timeospray: gravity_spheres_volume/dim_512/pathtracer/real_timecpuminer-opt: Blake-2 Scpuminer-opt: Garlicoincpuminer-opt: Skeincoincpuminer-opt: Myriad-Groestlcpuminer-opt: LBC, LBRY Creditscpuminer-opt: Quad SHA-256, Pyritelczero: Eigenonednn: IP Shapes 1D - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUospray-studio: 1 - 4K - 1 - Path Tracerospray-studio: 3 - 4K - 1 - Path Tracerospray-studio: 1 - 4K - 16 - Path Tracerospray-studio: 1 - 4K - 32 - Path Tracerospray-studio: 3 - 4K - 16 - Path Tracerospray-studio: 3 - 4K - 32 - Path Tracermnn: SqueezeNetV1.0ncnn: CPU - vision_transformeropenvino: Face Detection FP16 - CPUopenvino: Person Detection FP16 - CPUopenvino: Person Detection FP32 - CPUopenvino: Vehicle Detection FP16 - CPUopenvino: Face Detection FP16-INT8 - CPUopenvino: Vehicle Detection FP16-INT8 - CPUopenvino: Weld Porosity Detection FP16 - CPUopenvino: Machine Translation EN To DE FP16 - CPUopenvino: Weld Porosity Detection FP16-INT8 - CPUopenvino: Person Vehicle Bike Detection FP16 - CPUopenvino: Age Gender Recognition Retail 0013 FP16 - CPUopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUDefault, AVX-512 EnabledWithout AVX-512393.401451.9414.507.577.55742.9028.441890.251455.45134.592938.131696.8264463.9445555.6735.229335.12179.961.799.6410.266.01199250.6558.053977.934329.5090621134343977.802846675995715172032150317350.4171965.321270.4159400.5764321137.29583.3161137.96582.9013659436862029120727736171439813.54339.57549.941050.371054.9210.76280.864.235.4959.395.444.710.350.24380.321412.276.744.204.20502.2813.60953.79681.3766.031353.84907.3638316.8722781.7029.117030.38707.901.577.697.744.95153231.3784.535294.333986.9222710337002597.65204320163937696019266715780.6259427.449360.8001161.083961566.09910.4601544.39902.1024502534874844146156883441728723.99762.741178.811889.401887.1315.92585.818.3811.73121.0611.818.800.70.41OpenBenchmarking.org

CPU Temperature Monitor

OpenBenchmarking.orgCelsiusCPU Temperature MonitorPhoronix Test Suite System MonitoringDefault, AVX-512 EnabledWithout AVX-51220406080100Min: 33.63 / Avg: 81.24 / Max: 96.25Min: 36.13 / Avg: 84.12 / Max: 96.75

CPU Peak Freq (Highest CPU Core Frequency) Monitor

OpenBenchmarking.orgMegahertzCPU Peak Freq (Highest CPU Core Frequency) MonitorPhoronix Test Suite System MonitoringWithout AVX-512Default, AVX-512 Enabled10002000300040005000Min: 4495 / Avg: 5334.93 / Max: 5881Min: 4421 / Avg: 5372.19 / Max: 5881

CPU Power Consumption Monitor

OpenBenchmarking.orgWattsCPU Power Consumption MonitorPhoronix Test Suite System MonitoringDefault, AVX-512 EnabledWithout AVX-5124080120160200Min: 11.53 / Avg: 158.06 / Max: 237.76Min: 10.2 / Avg: 161.11 / Max: 235.88

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 1.0Video Input: Summer Nature 4KDefault, AVX-512 EnabledWithout AVX-51290180270360450SE +/- 1.37, N = 5SE +/- 3.22, N = 5393.40380.32-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi -lm-mno-avx512f1. (CC) gcc options: -O3 -march=native -pthread

OpenBenchmarking.orgFPS, More Is Betterdav1d 1.0Video Input: Summer Nature 1080pDefault, AVX-512 EnabledWithout AVX-51230060090012001500SE +/- 1.72, N = 10SE +/- 2.53, N = 101451.941412.27-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi -lm-mno-avx512f1. (CC) gcc options: -O3 -march=native -pthread

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Face Detection FP16 - Device: CPUDefault, AVX-512 EnabledWithout AVX-51248121620SE +/- 0.01, N = 3SE +/- 0.01, N = 314.506.74-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Person Detection FP16 - Device: CPUDefault, AVX-512 EnabledWithout AVX-512246810SE +/- 0.02, N = 3SE +/- 0.00, N = 37.574.20-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Person Detection FP32 - Device: CPUDefault, AVX-512 EnabledWithout AVX-512246810SE +/- 0.02, N = 3SE +/- 0.02, N = 37.554.20-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16 - Device: CPUDefault, AVX-512 EnabledWithout AVX-512160320480640800SE +/- 8.04, N = 3SE +/- 5.20, N = 3742.90502.28-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Face Detection FP16-INT8 - Device: CPUDefault, AVX-512 EnabledWithout AVX-512714212835SE +/- 0.01, N = 3SE +/- 0.02, N = 328.4413.60-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16-INT8 - Device: CPUDefault, AVX-512 EnabledWithout AVX-512400800120016002000SE +/- 2.10, N = 3SE +/- 0.65, N = 31890.25953.79-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16 - Device: CPUDefault, AVX-512 EnabledWithout AVX-51230060090012001500SE +/- 1.06, N = 3SE +/- 0.52, N = 31455.45681.37-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Machine Translation EN To DE FP16 - Device: CPUDefault, AVX-512 EnabledWithout AVX-512306090120150SE +/- 0.42, N = 3SE +/- 0.12, N = 3134.5966.03-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16-INT8 - Device: CPUDefault, AVX-512 EnabledWithout AVX-5126001200180024003000SE +/- 4.82, N = 3SE +/- 2.11, N = 32938.131353.84-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Person Vehicle Bike Detection FP16 - Device: CPUDefault, AVX-512 EnabledWithout AVX-512400800120016002000SE +/- 3.05, N = 3SE +/- 1.39, N = 31696.82907.36-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUDefault, AVX-512 EnabledWithout AVX-51214K28K42K56K70KSE +/- 24.16, N = 3SE +/- 43.53, N = 364463.9438316.87-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16 - Device: CPUDefault, AVX-512 EnabledWithout AVX-51210K20K30K40K50KSE +/- 43.30, N = 3SE +/- 51.81, N = 345555.6722781.70-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.13Binary: Pathtracer ISPC - Model: CrownDefault, AVX-512 EnabledWithout AVX-512816243240SE +/- 0.08, N = 3SE +/- 0.11, N = 335.2329.12MIN: 34.75 / MAX: 36.08MIN: 28.7 / MAX: 29.86

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.13Binary: Pathtracer ISPC - Model: Asian DragonDefault, AVX-512 EnabledWithout AVX-512816243240SE +/- 0.08, N = 3SE +/- 0.16, N = 335.1230.39MIN: 34.64 / MAX: 36.13MIN: 29.85 / MAX: 30.94

simdjson

This is a benchmark of SIMDJSON, a high performance JSON parser. SIMDJSON aims to be the fastest JSON parser and is used by projects like Microsoft FishStore, Yandex ClickHouse, Shopify, and others. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 2.0Throughput Test: TopTweetDefault, AVX-512 EnabledWithout AVX-5123691215SE +/- 0.09, N = 3SE +/- 0.02, N = 39.967.90-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native

OpenBenchmarking.orgGB/s Per Watt, More Is Bettersimdjson 2.0Throughput Test: KostyaDefault, AVX-512 EnabledWithout AVX-5120.02120.04240.06360.08480.1060.0940.076

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 2.0Throughput Test: LargeRandomDefault, AVX-512 EnabledWithout AVX-5120.40280.80561.20841.61122.014SE +/- 0.00, N = 3SE +/- 0.00, N = 31.791.57-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 2.0Throughput Test: PartialTweetsDefault, AVX-512 EnabledWithout AVX-5123691215SE +/- 0.08, N = 3SE +/- 0.01, N = 39.647.69-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 2.0Throughput Test: DistinctUserIDDefault, AVX-512 EnabledWithout AVX-5123691215SE +/- 0.06, N = 3SE +/- 0.07, N = 310.267.74-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 2.0Throughput Test: KostyaDefault, AVX-512 EnabledWithout AVX-512246810SE +/- 0.04, N = 3SE +/- 0.00, N = 36.014.95-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native

OpenVKL

OpenBenchmarking.orgItems / Sec Per Watt, More Is BetterOpenVKL 1.0Benchmark: vklBenchmark ISPCDefault, AVX-512 EnabledWithout AVX-5120.28730.57460.86191.14921.43651.2770.912

OpenBenchmarking.orgItems / Sec, More Is BetterOpenVKL 1.0Benchmark: vklBenchmark ISPCDefault, AVX-512 EnabledWithout AVX-5124080120160200SE +/- 0.33, N = 3SE +/- 0.00, N = 3199153MIN: 17 / MAX: 2254MIN: 14 / MAX: 1843

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.10Benchmark: particle_volume/pathtracer/real_timeDefault, AVX-512 EnabledWithout AVX-51250100150200250SE +/- 0.91, N = 3SE +/- 0.59, N = 3250.66231.38

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.10Benchmark: gravity_spheres_volume/dim_512/ao/real_timeDefault, AVX-512 EnabledWithout AVX-512246810SE +/- 0.00437, N = 3SE +/- 0.00574, N = 38.053974.53529

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.10Benchmark: gravity_spheres_volume/dim_512/scivis/real_timeDefault, AVX-512 EnabledWithout AVX-512246810SE +/- 0.03212, N = 3SE +/- 0.00159, N = 37.934324.33398

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.10Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_timeDefault, AVX-512 EnabledWithout AVX-5123691215SE +/- 0.00541, N = 3SE +/- 0.00119, N = 39.509066.92227

Cpuminer-Opt

OpenBenchmarking.orgkH/s Per Watt, More Is BetterCpuminer-Opt 3.18Algorithm: GarlicoinDefault, AVX-512 EnabledWithout AVX-51261218243024.6315.45

OpenBenchmarking.orgkH/s Per Watt, More Is BetterCpuminer-Opt 3.18Algorithm: SkeincoinDefault, AVX-512 EnabledWithout AVX-51260012001800240030002696.441926.96

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.18Algorithm: Blake-2 SDefault, AVX-512 EnabledWithout AVX-512500K1000K1500K2000K2500KSE +/- 38034.65, N = 12SE +/- 4192.40, N = 321134341033700-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native -lcurl -lz -lpthread -lssl -lcrypto -lgmp

OpenBenchmarking.orgkH/s Per Watt, More Is BetterCpuminer-Opt 3.18Algorithm: Myriad-GroestlDefault, AVX-512 EnabledWithout AVX-512110220330440550496.89128.78

OpenBenchmarking.orgkH/s Per Watt, More Is BetterCpuminer-Opt 3.18Algorithm: LBC, LBRY CreditsDefault, AVX-512 EnabledWithout AVX-512300600900120015001370.92579.30

OpenBenchmarking.orgkH/s Per Watt, More Is BetterCpuminer-Opt 3.18Algorithm: Quad SHA-256, PyriteDefault, AVX-512 EnabledWithout AVX-51260012001800240030003025.641788.77

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.18Algorithm: GarlicoinDefault, AVX-512 EnabledWithout AVX-5129001800270036004500SE +/- 75.31, N = 12SE +/- 25.96, N = 33977.802597.65-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native -lcurl -lz -lpthread -lssl -lcrypto -lgmp

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.18Algorithm: SkeincoinDefault, AVX-512 EnabledWithout AVX-51260K120K180K240K300KSE +/- 424.91, N = 3SE +/- 1803.62, N = 3284667204320-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native -lcurl -lz -lpthread -lssl -lcrypto -lgmp

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.18Algorithm: Myriad-GroestlDefault, AVX-512 EnabledWithout AVX-51213K26K39K52K65KSE +/- 669.95, N = 15SE +/- 105.88, N = 35995716393-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native -lcurl -lz -lpthread -lssl -lcrypto -lgmp

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.18Algorithm: LBC, LBRY CreditsDefault, AVX-512 EnabledWithout AVX-51230K60K90K120K150KSE +/- 177.76, N = 3SE +/- 40.41, N = 315172076960-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native -lcurl -lz -lpthread -lssl -lcrypto -lgmp

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 3.18Algorithm: Quad SHA-256, PyriteDefault, AVX-512 EnabledWithout AVX-51270K140K210K280K350KSE +/- 1040.42, N = 3SE +/- 177.04, N = 3321503192667-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native -lcurl -lz -lpthread -lssl -lcrypto -lgmp

LeelaChessZero

MinAvgMaxDefault, AVX-512 Enabled519153215881Without AVX-512515253075881OpenBenchmarking.orgMegahertz, More Is BetterLeelaChessZero 0.28CPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

oneDNN

MinAvgMaxDefault, AVX-512 Enabled512754195881Without AVX-512512553675881OpenBenchmarking.orgMegahertz, More Is BetteroneDNN 2.6CPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxWithout AVX-512531855855881Default, AVX-512 Enabled491054755881OpenBenchmarking.orgMegahertz, More Is BetteroneDNN 2.6CPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled518053815881Without AVX-512499253485881OpenBenchmarking.orgMegahertz, More Is BetteroneDNN 2.6CPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled520856135881Without AVX-512515755855881OpenBenchmarking.orgMegahertz, More Is BetteroneDNN 2.6CPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled482251665881Without AVX-512477250075881OpenBenchmarking.orgMegahertz, More Is BetteroneDNN 2.6CPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled480250915881Without AVX-512480450515881OpenBenchmarking.orgMegahertz, More Is BetteroneDNN 2.6CPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled456051635881Without AVX-512474549855881OpenBenchmarking.orgMegahertz, More Is BetteroneDNN 2.6CPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled442150885881Without AVX-512449550585881OpenBenchmarking.orgMegahertz, More Is BetteroneDNN 2.6CPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

OSPRay Studio

MinAvgMaxDefault, AVX-512 Enabled513652305881Without AVX-512512352125881OpenBenchmarking.orgMegahertz, More Is BetterOSPRay Studio 0.11CPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled513452275881Without AVX-512511752075881OpenBenchmarking.orgMegahertz, More Is BetterOSPRay Studio 0.11CPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled514252955881Without AVX-512511852545881OpenBenchmarking.orgMegahertz, More Is BetterOSPRay Studio 0.11CPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled511852295881Without AVX-512510652025881OpenBenchmarking.orgMegahertz, More Is BetterOSPRay Studio 0.11CPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled512852645881Without AVX-512512252405881OpenBenchmarking.orgMegahertz, More Is BetterOSPRay Studio 0.11CPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled513152205881Without AVX-512511851945881OpenBenchmarking.orgMegahertz, More Is BetterOSPRay Studio 0.11CPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

OpenVINO

MinAvgMaxDefault, AVX-512 Enabled503851805881Without AVX-512498851515881OpenBenchmarking.orgMegahertz, More Is BetterOpenVINO 2022.2.devCPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled503552295881Without AVX-512494751895881OpenBenchmarking.orgMegahertz, More Is BetterOpenVINO 2022.2.devCPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled505952195881Without AVX-512491751575881OpenBenchmarking.orgMegahertz, More Is BetterOpenVINO 2022.2.devCPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled514452655881Without AVX-512506251735881OpenBenchmarking.orgMegahertz, More Is BetterOpenVINO 2022.2.devCPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled506951885881Without AVX-512496351015881OpenBenchmarking.orgMegahertz, More Is BetterOpenVINO 2022.2.devCPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled506651475881Without AVX-512502751255881OpenBenchmarking.orgMegahertz, More Is BetterOpenVINO 2022.2.devCPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled507551595881Without AVX-512500851035881OpenBenchmarking.orgMegahertz, More Is BetterOpenVINO 2022.2.devCPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled513952435881Without AVX-512488851405881OpenBenchmarking.orgMegahertz, More Is BetterOpenVINO 2022.2.devCPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled458251635881Without AVX-512498250935881OpenBenchmarking.orgMegahertz, More Is BetterOpenVINO 2022.2.devCPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled510051885881Without AVX-512489951145881OpenBenchmarking.orgMegahertz, More Is BetterOpenVINO 2022.2.devCPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled500651075881Without AVX-512494550535881OpenBenchmarking.orgMegahertz, More Is BetterOpenVINO 2022.2.devCPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

MinAvgMaxDefault, AVX-512 Enabled506251425881Without AVX-512499951145881OpenBenchmarking.orgMegahertz, More Is BetterOpenVINO 2022.2.devCPU Peak Freq (Highest CPU Core Frequency) Monitor16003200480064008000

LeelaChessZero

LeelaChessZero (lc0 / lczero) is a chess engine automated vian neural networks. This test profile can be used for OpenCL, CUDA + cuDNN, and BLAS (CPU-based) benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgNodes Per Second, More Is BetterLeelaChessZero 0.28Backend: EigenDefault, AVX-512 EnabledWithout AVX-512400800120016002000SE +/- 7.31, N = 3SE +/- 16.33, N = 317351578-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -flto -O3 -march=native -pthread

oneDNN

MinAvgMaxDefault, AVX-512 Enabled48.470.887.5Without AVX-51248.180.293.6OpenBenchmarking.orgCelsius, Fewer Is BetteroneDNN 2.6CPU Temperature Monitor20406080100

MinAvgMaxWithout AVX-51245.158.865.6Default, AVX-512 Enabled42.069.782.9OpenBenchmarking.orgCelsius, Fewer Is BetteroneDNN 2.6CPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled41.375.886.6Without AVX-51240.981.994.9OpenBenchmarking.orgCelsius, Fewer Is BetteroneDNN 2.6CPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled43.572.482.6Without AVX-51244.877.991.6OpenBenchmarking.orgCelsius, Fewer Is BetteroneDNN 2.6CPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled42.587.892.0Without AVX-51243.588.195.9OpenBenchmarking.orgCelsius, Fewer Is BetteroneDNN 2.6CPU Temperature Monitor20406080100

MinAvgMaxWithout AVX-51247.586.495.8Default, AVX-512 Enabled47.089.292.6OpenBenchmarking.orgCelsius, Fewer Is BetteroneDNN 2.6CPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled47.588.892.0Without AVX-51247.389.896.1OpenBenchmarking.orgCelsius, Fewer Is BetteroneDNN 2.6CPU Temperature Monitor20406080100

MinAvgMaxWithout AVX-51247.686.896.0Default, AVX-512 Enabled47.588.992.8OpenBenchmarking.orgCelsius, Fewer Is BetteroneDNN 2.6CPU Temperature Monitor20406080100

OSPRay Studio

MinAvgMaxDefault, AVX-512 Enabled44.989.592.5Without AVX-51241.991.094.1OpenBenchmarking.orgCelsius, Fewer Is BetterOSPRay Studio 0.11CPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled47.089.792.5Without AVX-51247.691.894.5OpenBenchmarking.orgCelsius, Fewer Is BetterOSPRay Studio 0.11CPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled46.987.892.4Without AVX-51248.090.295.0OpenBenchmarking.orgCelsius, Fewer Is BetterOSPRay Studio 0.11CPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled46.690.093.5Without AVX-51248.192.395.3OpenBenchmarking.orgCelsius, Fewer Is BetterOSPRay Studio 0.11CPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled47.588.393.0Without AVX-51248.390.494.5OpenBenchmarking.orgCelsius, Fewer Is BetterOSPRay Studio 0.11CPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled47.490.093.4Without AVX-51248.092.395.1OpenBenchmarking.orgCelsius, Fewer Is BetterOSPRay Studio 0.11CPU Temperature Monitor20406080100

OpenVINO

MinAvgMaxWithout AVX-51241.893.096.3Default, AVX-512 Enabled40.993.195.9OpenBenchmarking.orgCelsius, Fewer Is BetterOpenVINO 2022.2.devCPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled47.193.296.1Without AVX-51247.694.096.6OpenBenchmarking.orgCelsius, Fewer Is BetterOpenVINO 2022.2.devCPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled47.393.396.3Without AVX-51248.494.196.8OpenBenchmarking.orgCelsius, Fewer Is BetterOpenVINO 2022.2.devCPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled47.583.486.0Without AVX-51247.991.294.1OpenBenchmarking.orgCelsius, Fewer Is BetterOpenVINO 2022.2.devCPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled45.893.296.0Without AVX-51248.393.395.6OpenBenchmarking.orgCelsius, Fewer Is BetterOpenVINO 2022.2.devCPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled47.393.395.4Without AVX-51248.393.595.4OpenBenchmarking.orgCelsius, Fewer Is BetterOpenVINO 2022.2.devCPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled47.893.395.4Without AVX-51248.093.995.8OpenBenchmarking.orgCelsius, Fewer Is BetterOpenVINO 2022.2.devCPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled47.387.290.1Without AVX-51248.693.796.8OpenBenchmarking.orgCelsius, Fewer Is BetterOpenVINO 2022.2.devCPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled45.993.195.3Without AVX-51247.993.595.4OpenBenchmarking.orgCelsius, Fewer Is BetterOpenVINO 2022.2.devCPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled47.591.194.1Without AVX-51248.393.496.0OpenBenchmarking.orgCelsius, Fewer Is BetterOpenVINO 2022.2.devCPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled47.093.295.5Without AVX-51248.693.295.5OpenBenchmarking.orgCelsius, Fewer Is BetterOpenVINO 2022.2.devCPU Temperature Monitor20406080100

MinAvgMaxDefault, AVX-512 Enabled48.093.295.5Without AVX-51248.593.795.6OpenBenchmarking.orgCelsius, Fewer Is BetterOpenVINO 2022.2.devCPU Temperature Monitor20406080100

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUDefault, AVX-512 EnabledWithout AVX-5120.14080.28160.42240.56320.704SE +/- 0.007969, N = 15SE +/- 0.006181, N = 50.4171960.625942-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 0.34-mno-avx512f - MIN: 0.581. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUDefault, AVX-512 EnabledWithout AVX-512246810SE +/- 0.00255, N = 7SE +/- 0.00637, N = 75.321277.44936-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 5.26-mno-avx512f - MIN: 7.291. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUDefault, AVX-512 EnabledWithout AVX-5120.180.360.540.720.9SE +/- 0.000107, N = 3SE +/- 0.010456, N = 30.4159400.800116-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 0.4-mno-avx512f - MIN: 0.761. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUDefault, AVX-512 EnabledWithout AVX-5120.24390.48780.73170.97561.2195SE +/- 0.001581, N = 9SE +/- 0.006665, N = 90.5764321.083960-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 0.56-mno-avx512f - MIN: 1.031. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUDefault, AVX-512 EnabledWithout AVX-51230060090012001500SE +/- 1.54, N = 3SE +/- 8.37, N = 31137.291566.09-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 1130.88-mno-avx512f - MIN: 1541.571. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUDefault, AVX-512 EnabledWithout AVX-5122004006008001000SE +/- 0.21, N = 3SE +/- 3.36, N = 3583.32910.46-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 579.29-mno-avx512f - MIN: 898.051. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUDefault, AVX-512 EnabledWithout AVX-51230060090012001500SE +/- 1.66, N = 3SE +/- 19.60, N = 31137.961544.39-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 1131.23-mno-avx512f - MIN: 1494.511. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUDefault, AVX-512 EnabledWithout AVX-5122004006008001000SE +/- 1.02, N = 3SE +/- 5.88, N = 3582.90902.10-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 578.18-mno-avx512f - MIN: 885.671. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl -lpthread

OSPRay Studio

Intel OSPRay Studio is an open-source, interactive visualization and ray-tracing software package. OSPRay Studio makes use of Intel OSPRay, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 1 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path TracerDefault, AVX-512 EnabledWithout AVX-51210002000300040005000SE +/- 5.21, N = 3SE +/- 6.51, N = 336594502-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native -ldl

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 3 - Resolution: 4K - Samples Per Pixel: 1 - Renderer: Path TracerDefault, AVX-512 EnabledWithout AVX-51211002200330044005500SE +/- 5.55, N = 3SE +/- 3.28, N = 343685348-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native -ldl

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 1 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path TracerDefault, AVX-512 EnabledWithout AVX-51216K32K48K64K80KSE +/- 138.54, N = 3SE +/- 91.34, N = 36202974844-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native -ldl

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 1 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path TracerDefault, AVX-512 EnabledWithout AVX-51230K60K90K120K150KSE +/- 140.99, N = 3SE +/- 168.25, N = 3120727146156-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native -ldl

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 3 - Resolution: 4K - Samples Per Pixel: 16 - Renderer: Path TracerDefault, AVX-512 EnabledWithout AVX-51220K40K60K80K100KSE +/- 149.41, N = 3SE +/- 7.57, N = 37361788344-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native -ldl

OpenBenchmarking.orgms, Fewer Is BetterOSPRay Studio 0.11Camera: 3 - Resolution: 4K - Samples Per Pixel: 32 - Renderer: Path TracerDefault, AVX-512 EnabledWithout AVX-51240K80K120K160K200KSE +/- 116.66, N = 3SE +/- 155.59, N = 3143981172872-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi-mno-avx512f1. (CXX) g++ options: -O3 -march=native -ldl

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. This MNN test profile is building the OpenMP / CPU threaded version for processor benchmarking and not any GPU-accelerated test. MNN does allow making use of AVX-512 extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: SqueezeNetV1.0Default, AVX-512 EnabledWithout AVX-5120.89931.79862.69793.59724.4965SE +/- 0.015, N = 6SE +/- 0.121, N = 33.5433.997-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 3.43 / MAX: 10.23-mno-avx512f - MIN: 3.72 / MAX: 5.691. (CXX) g++ options: -O3 -march=native -std=c++11 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20220729Target: CPU - Model: vision_transformerDefault, AVX-512 EnabledWithout AVX-5121428425670SE +/- 0.03, N = 3SE +/- 0.91, N = 339.5762.74-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 39.32 / MAX: 41.99-mno-avx512f - MIN: 61.57 / MAX: 65.211. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Face Detection FP16 - Device: CPUDefault, AVX-512 EnabledWithout AVX-51230060090012001500SE +/- 0.48, N = 3SE +/- 0.92, N = 3549.941178.81-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 282.71 / MAX: 577.01-mno-avx512f - MIN: 596.16 / MAX: 1250.721. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Person Detection FP16 - Device: CPUDefault, AVX-512 EnabledWithout AVX-512400800120016002000SE +/- 2.40, N = 3SE +/- 0.47, N = 31050.371889.40-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 604.91 / MAX: 1254.32-mno-avx512f - MIN: 1032.21 / MAX: 2215.281. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Person Detection FP32 - Device: CPUDefault, AVX-512 EnabledWithout AVX-512400800120016002000SE +/- 3.05, N = 3SE +/- 5.93, N = 31054.921887.13-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 641.2 / MAX: 1276.51-mno-avx512f - MIN: 1054.92 / MAX: 2161.51. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16 - Device: CPUDefault, AVX-512 EnabledWithout AVX-51248121620SE +/- 0.12, N = 3SE +/- 0.16, N = 310.7615.92-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 4.25 / MAX: 23.39-mno-avx512f - MIN: 6.75 / MAX: 34.351. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Face Detection FP16-INT8 - Device: CPUDefault, AVX-512 EnabledWithout AVX-512130260390520650SE +/- 0.17, N = 3SE +/- 0.73, N = 3280.86585.81-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 263.49 / MAX: 294.78-mno-avx512f - MIN: 556.1 / MAX: 5961. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16-INT8 - Device: CPUDefault, AVX-512 EnabledWithout AVX-512246810SE +/- 0.01, N = 3SE +/- 0.01, N = 34.238.38-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 2.63 / MAX: 12.45-mno-avx512f - MIN: 4.65 / MAX: 17.131. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16 - Device: CPUDefault, AVX-512 EnabledWithout AVX-5123691215SE +/- 0.00, N = 3SE +/- 0.01, N = 35.4911.73-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 2.89 / MAX: 13.22-mno-avx512f - MIN: 6.24 / MAX: 20.451. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Machine Translation EN To DE FP16 - Device: CPUDefault, AVX-512 EnabledWithout AVX-512306090120150SE +/- 0.18, N = 3SE +/- 0.21, N = 359.39121.06-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 26.85 / MAX: 70.78-mno-avx512f - MIN: 57.76 / MAX: 141.81. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16-INT8 - Device: CPUDefault, AVX-512 EnabledWithout AVX-5123691215SE +/- 0.01, N = 3SE +/- 0.02, N = 35.4411.81-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 2.88 / MAX: 13.32-mno-avx512f - MIN: 6.9 / MAX: 19.21. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Person Vehicle Bike Detection FP16 - Device: CPUDefault, AVX-512 EnabledWithout AVX-512246810SE +/- 0.01, N = 3SE +/- 0.01, N = 34.718.80-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 3.46 / MAX: 12.8-mno-avx512f - MIN: 5.34 / MAX: 19.181. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16 - Device: CPUDefault, AVX-512 EnabledWithout AVX-5120.15750.3150.47250.630.7875SE +/- 0.00, N = 3SE +/- 0.00, N = 30.350.70-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 0.21 / MAX: 7.78-mno-avx512f - MIN: 0.39 / MAX: 8.821. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUDefault, AVX-512 EnabledWithout AVX-5120.09230.18460.27690.36920.4615SE +/- 0.00, N = 3SE +/- 0.00, N = 30.240.41-mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi - MIN: 0.15 / MAX: 7.55-mno-avx512f - MIN: 0.22 / MAX: 7.871. (CXX) g++ options: -fPIC -O3 -march=native -fsigned-char -ffunction-sections -fdata-sections -fno-strict-overflow -fwrapv -flto -shared

124 Results Shown

CPU Temperature Monitor:
  Phoronix Test Suite System Monitoring:
    Celsius
    Megahertz
    Watts
dav1d:
  Summer Nature 4K
  Summer Nature 1080p
OpenVINO:
  Face Detection FP16 - CPU
  Person Detection FP16 - CPU
  Person Detection FP32 - CPU
  Vehicle Detection FP16 - CPU
  Face Detection FP16-INT8 - CPU
  Vehicle Detection FP16-INT8 - CPU
  Weld Porosity Detection FP16 - CPU
  Machine Translation EN To DE FP16 - CPU
  Weld Porosity Detection FP16-INT8 - CPU
  Person Vehicle Bike Detection FP16 - CPU
  Age Gender Recognition Retail 0013 FP16-INT8 - CPU
  Age Gender Recognition Retail 0013 FP16 - CPU
Embree:
  Pathtracer ISPC - Crown
  Pathtracer ISPC - Asian Dragon
simdjson
simdjson
simdjson:
  LargeRand
  PartialTweets
  DistinctUserID
  Kostya
OpenVKL
OpenVKL
OSPRay:
  particle_volume/pathtracer/real_time
  gravity_spheres_volume/dim_512/ao/real_time
  gravity_spheres_volume/dim_512/scivis/real_time
  gravity_spheres_volume/dim_512/pathtracer/real_time
Cpuminer-Opt:
  Garlicoin
  Skeincoin
Cpuminer-Opt
Cpuminer-Opt:
  Myriad-Groestl
  LBC, LBRY Credits
  Quad SHA-256, Pyrite
Cpuminer-Opt:
  Garlicoin
  Skeincoin
  Myriad-Groestl
  LBC, LBRY Credits
  Quad SHA-256, Pyrite
LeelaChessZero:
  CPU Peak Freq (Highest CPU Core Frequency) Monitor:
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
    Megahertz
LeelaChessZero
oneDNN:
  CPU Temp Monitor:
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
    Celsius
oneDNN:
  IP Shapes 1D - u8s8f32 - CPU
  Convolution Batch Shapes Auto - u8s8f32 - CPU
  Deconvolution Batch shapes_1d - u8s8f32 - CPU
  Deconvolution Batch shapes_3d - u8s8f32 - CPU
  Recurrent Neural Network Training - u8s8f32 - CPU
  Recurrent Neural Network Inference - u8s8f32 - CPU
  Recurrent Neural Network Training - bf16bf16bf16 - CPU
  Recurrent Neural Network Inference - bf16bf16bf16 - CPU
OSPRay Studio:
  1 - 4K - 1 - Path Tracer
  3 - 4K - 1 - Path Tracer
  1 - 4K - 16 - Path Tracer
  1 - 4K - 32 - Path Tracer
  3 - 4K - 16 - Path Tracer
  3 - 4K - 32 - Path Tracer
Mobile Neural Network
NCNN
OpenVINO:
  Face Detection FP16 - CPU
  Person Detection FP16 - CPU
  Person Detection FP32 - CPU
  Vehicle Detection FP16 - CPU
  Face Detection FP16-INT8 - CPU
  Vehicle Detection FP16-INT8 - CPU
  Weld Porosity Detection FP16 - CPU
  Machine Translation EN To DE FP16 - CPU
  Weld Porosity Detection FP16-INT8 - CPU
  Person Vehicle Bike Detection FP16 - CPU
  Age Gender Recognition Retail 0013 FP16 - CPU
  Age Gender Recognition Retail 0013 FP16-INT8 - CPU