AMD 3D V-Cache Comparison

Tests for a future article.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2204297-NE-CC929132156
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results
Show Result Confidence Charts
Allow Limiting Results To Certain Suite(s)

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Disable Color Branding
Prefer Vertical Bar Graphs

Additional Graphs

Show Perf Per Core/Thread Calculation Graphs Where Applicable
Show Perf Per Clock Calculation Graphs Where Applicable

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Toggle/Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
Ryzen 9 5950X
April 26 2022
  14 Hours, 53 Minutes
Ryzen 7 5800X3D
April 26 2022
  14 Hours, 56 Minutes
Ryzen 7 5800X
April 27 2022
  18 Hours, 29 Minutes
Ryzen 9 5900X
April 28 2022
  14 Hours, 16 Minutes
Core i9 12900K
April 28 2022
  14 Hours, 51 Minutes
Invert Behavior (Only Show Selected Data)
  15 Hours, 29 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


AMD 3D V-Cache ComparisonProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerOpenGLVulkanCompilerFile-SystemScreen ResolutionDisplay DriverRyzen 9 5950XRyzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XCore i9 12900KAMD Ryzen 9 5950X 16-Core @ 3.40GHz (16 Cores / 32 Threads)ASUS ROG CROSSHAIR VIII HERO (WI-FI) (4006 BIOS)AMD Starship/Matisse32GB1000GB Sabrent Rocket 4.0 1TBAMD Radeon RX 6800 16GB (2475/1000MHz)AMD Navi 21 HDMI AudioASUS MG28URealtek RTL8125 2.5GbE + Intel I211 + Intel Wi-Fi 6 AX200Ubuntu 22.045.17.4-051704-generic (x86_64)GNOME Shell 42.0X Server + Wayland4.6 Mesa 22.2.0-devel (git-092ac67 2022-04-21 jammy-oibaf-ppa) (LLVM 14.0.0 DRM 3.44)1.3.211GCC 11.2.0ext43840x2160AMD Ryzen 7 5800X3D 8-Core @ 3.40GHz (8 Cores / 16 Threads)ASRock X570 Pro4 (P4.30 BIOS)16GBAMD Radeon RX 6800 XT 16GB (2575/1000MHz)ASUS VP28UIntel I211AMD Ryzen 7 5800X 8-Core @ 3.80GHz (8 Cores / 16 Threads)AMD Ryzen 9 5900X 12-Core @ 3.70GHz (12 Cores / 24 Threads)ASUS ROG CROSSHAIR VIII HERO (3904 BIOS)NVIDIA NV134 8GBNVIDIA GP104 HD AudioASUS MG28URealtek RTL8125 2.5GbE + Intel I211nouveau4.3 Mesa 22.2.0-devel (git-092ac67 2022-04-21 jammy-oibaf-ppa)Intel Core i9-12900K @ 5.20GHz (16 Cores / 24 Threads)ASUS ROG STRIX Z690-E GAMING WIFI (1003 BIOS)Intel Device 7aa732GBAMD Radeon RX 6800 XT 16GB (2575/1000MHz)Intel Device 7ad0ASUS VP28UIntel I225-V + Intel Wi-Fi 6 AX210/AX211/AX4114.6 Mesa 22.2.0-devel (git-092ac67 2022-04-21 jammy-oibaf-ppa) (LLVM 14.0.0 DRM 3.44)OpenBenchmarking.orgKernel Details- Transparent Huge Pages: madviseCompiler Details- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- Ryzen 9 5950X: Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0xa201016- Ryzen 7 5800X3D: Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0xa201205- Ryzen 7 5800X: Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0xa201016- Ryzen 9 5900X: Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0xa201016- Core i9 12900K: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.9Python Details- Python 3.10.4Security Details- Ryzen 9 5950X: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: always-on RSB filling + srbds: Not affected + tsx_async_abort: Not affected- Ryzen 7 5800X3D: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: always-on RSB filling + srbds: Not affected + tsx_async_abort: Not affected- Ryzen 7 5800X: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: always-on RSB filling + srbds: Not affected + tsx_async_abort: Not affected- Ryzen 9 5900X: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: always-on RSB filling + srbds: Not affected + tsx_async_abort: Not affected- Core i9 12900K: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected

Ryzen 9 5950XRyzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XCore i9 12900KResult OverviewPhoronix Test Suite100%156%211%267%323%LeelaChessZeroXcompact3d Incompact3dOpenFOAMASKAPCaffeWebP2 Image EncodeTNNNCNNMlpack BenchmarkMobile Neural NetworkONNX RuntimeoneDNNNumpy BenchmarkECP-CANDLEOpen Porous Media Git

AMD 3D V-Cache Comparisonopenfoam: Motorbike 60Mopm-git: Flow MPI Extra - 1lczero: Eigenopm-git: Flow MPI Extra - 8opm-git: Flow MPI Extra - 2opm-git: Flow MPI Extra - 4webp2: Quality 100, Lossless Compressionopm-git: Flow MPI Norne-4C MSW - 8opm-git: Flow MPI Norne-4C MSW - 1opm-git: Flow MPI Norne-4C MSW - 2opm-git: Flow MPI Norne - 8opm-git: Flow MPI Norne-4C MSW - 4onnx: yolov4 - CPU - Standardlczero: BLASecp-candle: P3B1webp2: Quality 95, Compression Effort 7opm-git: Flow MPI Norne - 1opm-git: Flow MPI Norne - 4ecp-candle: P3B2onnx: ArcFace ResNet-100 - CPU - Standardonnx: GPT-2 - CPU - Standardopm-git: Flow MPI Norne - 2onnx: fcn-resnet101-11 - CPU - Standardonnx: bertsquad-12 - CPU - Standardonnx: super-resolution-10 - CPU - Standardncnn: CPU - regnety_400mncnn: CPU - squeezenet_ssdncnn: CPU - yolov4-tinyncnn: CPU - resnet50ncnn: CPU - alexnetncnn: CPU - resnet18ncnn: CPU - vgg16ncnn: CPU - googlenetncnn: CPU - blazefacencnn: CPU - efficientnet-b0ncnn: CPU - mnasnetncnn: CPU - shufflenet-v2ncnn: CPU-v3-v3 - mobilenet-v3ncnn: CPU-v2-v2 - mobilenet-v2ncnn: CPU - mobilenetincompact3d: input.i3d 193 Cells Per Directiontnn: CPU - DenseNetcaffe: GoogleNet - CPU - 200numpy: webp2: Quality 75, Compression Effort 7onnx: super-resolution-10 - CPU - Parallelonnx: fcn-resnet101-11 - CPU - Parallelonnx: GPT-2 - CPU - Parallelonnx: ArcFace ResNet-100 - CPU - Parallelonnx: yolov4 - CPU - Parallelonnx: bertsquad-12 - CPU - Parallelaskap: tConvolve MT - Degriddingaskap: tConvolve MT - Griddingopenfoam: Motorbike 30Mcaffe: GoogleNet - CPU - 100onednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUmlpack: scikit_qdaonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUmnn: inception-v3mnn: mobilenet-v1-1.0mnn: MobileNetV2_224mnn: SqueezeNetV1.0mnn: resnet-v2-50mnn: squeezenetv1.1mnn: mobilenetV3onednn: Deconvolution Batch shapes_1d - f32 - CPUmlpack: scikit_linearridgeregressioncaffe: AlexNet - CPU - 200askap: tConvolve MPI - Griddingaskap: tConvolve MPI - Degriddingmlpack: scikit_icacaffe: AlexNet - CPU - 100onednn: IP Shapes 1D - u8s8f32 - CPUincompact3d: input.i3d 129 Cells Per Directionaskap: tConvolve OpenMP - Degriddingaskap: tConvolve OpenMP - Griddingonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUonednn: IP Shapes 1D - f32 - CPUtnn: CPU - MobileNet v2mlpack: scikit_svmtnn: CPU - SqueezeNet v1.1askap: Hogbom Clean OpenMPonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: IP Shapes 3D - f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUwebp2: Quality 100, Compression Effort 5tnn: CPU - SqueezeNet v2ecp-candle: P1B2onednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUwebp2: DefaultRyzen 9 5950XRyzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XCore i9 12900K1382.541074.33684808.05712.62690.62536.767575.13569.14441.07357.86377.034876681309.594234.927270.81232.61654.35316687062203.899880061699.6114.5520.4924.4011.0814.2856.5512.921.805.233.874.163.774.2912.27156.1625262425.594193710591.54110.284643688553415402965541346.57784.50798.44966241820.402750.182703.332745.4455.061814.131783.8525.7262.6393.2755.23120.5153.9882.3914.755331.69733526728.096643.6441.77366580.82300933.94812143214.582755.721.069263.89647224.26416.56213.323220.03616.705118.03217.796520.4760973.24950.94031.3923.626091.618992.1651090.21997.341160781.98672.48643.96864.935581.46476.91403.35361.91357.3357212541023.315376.956224.10223.97553.74519398826187.4910782241075.1812.4414.8018.139.6210.2742.607.311.063.012.012.111.902.137.62126.8610332609.843162022603.09178.013460672691912233085181594.95930.57480.29808621385.792691.242683.162691.6738.271382.161387.8623.1931.6761.8314.21316.3562.5901.0827.276231.63604948746.526453.2234.88302221.2378527.18149768741.66946.111.776412.89717233.17916.17222.332527.72312.622710.38006.557650.6047805.16953.22129.9925.571092.483933.1701270.721154.098541027.17799.40829.49974.591733.91601.31511.31447.63454.344318671158.674420.032281.47275.40390.26910426832234.634955736285.9316.7619.7820.2211.8313.0255.9710.221.223.602.242.352.282.6111.63141.2109543016.479179146491.56199.645410367562111072734801471.58854.707177.60898251847.703065.263053.303055.8452.571859.141864.5425.6921.8161.9424.53618.4422.8051.1568.342031.82676614256.053976.3039.16339051.4072837.21053443367.671709.252.036003.29136272.03720.29266.187221.73218.753516.35348.935521.457525.87563.60534.8416.370232.839133.6191277.551073.71886810.33713.34691.38614.812581.35572.60441.14361.84378.335399541174.469268.612273.89233.63667.28215197862203.9811594975458.4313.5120.6321.2110.0112.5050.7511.441.634.773.443.883.463.9111.42144.4737502563.757178725605.17129.733559185568714212995401541.67837.94496.16894921776.192904.332873.452908.2851.391785.051742.0623.6924.0722.9754.75824.4993.3941.8385.386891.58686768258.027668.0539.37343291.00694132.15362933732.722936.851.300493.44548224.19916.15213.776238.66716.139816.73827.548430.5197333.70950.85230.2414.387731.997422.439487.801053.382161891.78717.04671.56476.451825.56537.38475.95515.71451.006932201429.61215.582247.89280.851503.264197511082223.2010198847477.3913.2715.8616.847.589.6628.249.941.465.343.103.102.903.4111.1155.60967131792.972134045665.34101.568451611180883626299193872.032721.1584.47678331617.102881.322881.722881.1027.311613.811616.0624.3012.8912.414.15123.0852.4051.1748.737501.60517134472.734198.5132.93255901.0520014.55304507793.774631.821.352122.63611170.58310.54133.970540.5415.875366.002143.446430.8120472.93638.91520.9415.253942.223362.062OpenBenchmarking.org

OpenFOAM

OpenFOAM is the leading free, open source software for computational fluid dynamics (CFD). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenFOAM 8Input: Motorbike 60MCore i9 12900KRyzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XRyzen 9 5950X30060090012001500SE +/- 0.14, N = 3SE +/- 1.12, N = 3SE +/- 0.23, N = 3SE +/- 0.77, N = 3SE +/- 0.26, N = 3487.801090.211270.721277.551382.54-lfoamToVTK -llagrangian -lfileFormats1. (CXX) g++ options: -std=c++11 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -ldynamicMesh -lgenericPatchFields -lOpenFOAM -ldl -lm

Open Porous Media Git

This is a test of Open Porous Media, a set of open-source tools concerning simulation of flow and transport of fluids in porous media. This test profile builds OPM and its dependencies from upstream Git. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpen Porous Media GitOPM Benchmark: Flow MPI Extra - Threads: 1Ryzen 7 5800X3DCore i9 12900KRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X2004006008001000SE +/- 2.03, N = 3SE +/- 2.12, N = 3SE +/- 4.75, N = 3SE +/- 3.75, N = 3SE +/- 5.46, N = 3997.341053.381073.711074.331154.091. (CXX) g++ options: -pipe -pthread -fopenmp -O3 -mtune=native -UNDEBUG -lm -ldl -lrt2. Ryzen 7 5800X3D: Build Time Mon Apr 25 06:10:54 PM EDT 20223. Core i9 12900K: Build Time Thu Apr 28 06:45:36 PM EDT 20224. Ryzen 9 5900X: Build Time Mon Apr 25 06:10:54 PM EDT 20225. Ryzen 9 5950X: Build Time Mon Apr 25 06:10:54 PM EDT 20226. Ryzen 7 5800X: Build Time Mon Apr 25 06:10:54 PM EDT 2022

LeelaChessZero

LeelaChessZero (lc0 / lczero) is a chess engine automated vian neural networks. This test profile can be used for OpenCL, CUDA + cuDNN, and BLAS (CPU-based) benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgNodes Per Second, More Is BetterLeelaChessZero 0.28Backend: EigenCore i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 7 5800XRyzen 9 5950X5001000150020002500SE +/- 19.84, N = 7SE +/- 8.89, N = 3SE +/- 7.53, N = 9SE +/- 8.85, N = 9SE +/- 7.32, N = 9216111608868546841. (CXX) g++ options: -flto -pthread

Open Porous Media Git

This is a test of Open Porous Media, a set of open-source tools concerning simulation of flow and transport of fluids in porous media. This test profile builds OPM and its dependencies from upstream Git. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpen Porous Media GitOPM Benchmark: Flow MPI Extra - Threads: 8Ryzen 7 5800X3DRyzen 9 5950XRyzen 9 5900XCore i9 12900KRyzen 7 5800X2004006008001000SE +/- 1.20, N = 3SE +/- 0.30, N = 3SE +/- 0.19, N = 3SE +/- 0.86, N = 3SE +/- 0.32, N = 3781.98808.05810.33891.781027.171. (CXX) g++ options: -pipe -pthread -fopenmp -O3 -mtune=native -UNDEBUG -lm -ldl -lrt2. Ryzen 7 5800X3D: Build Time Mon Apr 25 06:10:54 PM EDT 20223. Ryzen 9 5950X: Build Time Mon Apr 25 06:10:54 PM EDT 20224. Ryzen 9 5900X: Build Time Mon Apr 25 06:10:54 PM EDT 20225. Core i9 12900K: Build Time Thu Apr 28 06:45:36 PM EDT 20226. Ryzen 7 5800X: Build Time Mon Apr 25 06:10:54 PM EDT 2022

OpenBenchmarking.orgSeconds, Fewer Is BetterOpen Porous Media GitOPM Benchmark: Flow MPI Extra - Threads: 2Ryzen 7 5800X3DRyzen 9 5950XRyzen 9 5900XCore i9 12900KRyzen 7 5800X2004006008001000SE +/- 0.69, N = 3SE +/- 1.57, N = 3SE +/- 1.56, N = 3SE +/- 0.67, N = 3SE +/- 2.38, N = 3672.48712.62713.34717.04799.401. (CXX) g++ options: -pipe -pthread -fopenmp -O3 -mtune=native -UNDEBUG -lm -ldl -lrt2. Ryzen 7 5800X3D: Build Time Mon Apr 25 06:10:54 PM EDT 20223. Ryzen 9 5950X: Build Time Mon Apr 25 06:10:54 PM EDT 20224. Ryzen 9 5900X: Build Time Mon Apr 25 06:10:54 PM EDT 20225. Core i9 12900K: Build Time Thu Apr 28 06:45:36 PM EDT 20226. Ryzen 7 5800X: Build Time Mon Apr 25 06:10:54 PM EDT 2022

OpenBenchmarking.orgSeconds, Fewer Is BetterOpen Porous Media GitOPM Benchmark: Flow MPI Extra - Threads: 4Ryzen 7 5800X3DCore i9 12900KRyzen 9 5950XRyzen 9 5900XRyzen 7 5800X2004006008001000SE +/- 0.34, N = 3SE +/- 0.81, N = 3SE +/- 0.65, N = 3SE +/- 0.23, N = 3SE +/- 0.80, N = 3643.96671.56690.62691.38829.491. (CXX) g++ options: -pipe -pthread -fopenmp -O3 -mtune=native -UNDEBUG -lm -ldl -lrt2. Ryzen 7 5800X3D: Build Time Mon Apr 25 06:10:54 PM EDT 20223. Core i9 12900K: Build Time Thu Apr 28 06:45:36 PM EDT 20224. Ryzen 9 5950X: Build Time Mon Apr 25 06:10:54 PM EDT 20225. Ryzen 9 5900X: Build Time Mon Apr 25 06:10:54 PM EDT 20226. Ryzen 7 5800X: Build Time Mon Apr 25 06:10:54 PM EDT 2022

WebP2 Image Encode

This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterWebP2 Image Encode 20220422Encode Settings: Quality 100, Lossless CompressionCore i9 12900KRyzen 9 5950XRyzen 9 5900XRyzen 7 5800X3DRyzen 7 5800X2004006008001000SE +/- 0.96, N = 3SE +/- 2.59, N = 3SE +/- 4.10, N = 3SE +/- 2.17, N = 3SE +/- 1.21, N = 3476.45536.77614.81864.94974.591. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl

Open Porous Media Git

This is a test of Open Porous Media, a set of open-source tools concerning simulation of flow and transport of fluids in porous media. This test profile builds OPM and its dependencies from upstream Git. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpen Porous Media GitOPM Benchmark: Flow MPI Norne-4C MSW - Threads: 8Ryzen 9 5950XRyzen 9 5900XRyzen 7 5800X3DRyzen 7 5800XCore i9 12900K2004006008001000SE +/- 0.32, N = 3SE +/- 0.16, N = 3SE +/- 0.28, N = 3SE +/- 0.32, N = 3SE +/- 0.65, N = 3575.13581.35581.46733.91825.561. (CXX) g++ options: -pipe -pthread -fopenmp -O3 -mtune=native -UNDEBUG -lm -ldl -lrt2. Ryzen 9 5950X: Build Time Mon Apr 25 06:10:54 PM EDT 20223. Ryzen 9 5900X: Build Time Mon Apr 25 06:10:54 PM EDT 20224. Ryzen 7 5800X3D: Build Time Mon Apr 25 06:10:54 PM EDT 20225. Ryzen 7 5800X: Build Time Mon Apr 25 06:10:54 PM EDT 20226. Core i9 12900K: Build Time Thu Apr 28 06:45:36 PM EDT 2022

OpenBenchmarking.orgSeconds, Fewer Is BetterOpen Porous Media GitOPM Benchmark: Flow MPI Norne-4C MSW - Threads: 1Ryzen 7 5800X3DCore i9 12900KRyzen 9 5950XRyzen 9 5900XRyzen 7 5800X130260390520650SE +/- 0.27, N = 3SE +/- 1.75, N = 3SE +/- 2.73, N = 3SE +/- 0.96, N = 3SE +/- 1.44, N = 3476.91537.38569.14572.60601.311. (CXX) g++ options: -pipe -pthread -fopenmp -O3 -mtune=native -UNDEBUG -lm -ldl -lrt2. Ryzen 7 5800X3D: Build Time Mon Apr 25 06:10:54 PM EDT 20223. Core i9 12900K: Build Time Thu Apr 28 06:45:36 PM EDT 20224. Ryzen 9 5950X: Build Time Mon Apr 25 06:10:54 PM EDT 20225. Ryzen 9 5900X: Build Time Mon Apr 25 06:10:54 PM EDT 20226. Ryzen 7 5800X: Build Time Mon Apr 25 06:10:54 PM EDT 2022

OpenBenchmarking.orgSeconds, Fewer Is BetterOpen Porous Media GitOPM Benchmark: Flow MPI Norne-4C MSW - Threads: 2Ryzen 7 5800X3DRyzen 9 5950XRyzen 9 5900XCore i9 12900KRyzen 7 5800X110220330440550SE +/- 0.45, N = 3SE +/- 0.66, N = 3SE +/- 0.35, N = 3SE +/- 0.98, N = 3SE +/- 0.29, N = 3403.35441.07441.14475.95511.311. (CXX) g++ options: -pipe -pthread -fopenmp -O3 -mtune=native -UNDEBUG -lm -ldl -lrt2. Ryzen 7 5800X3D: Build Time Mon Apr 25 06:10:54 PM EDT 20223. Ryzen 9 5950X: Build Time Mon Apr 25 06:10:54 PM EDT 20224. Ryzen 9 5900X: Build Time Mon Apr 25 06:10:54 PM EDT 20225. Core i9 12900K: Build Time Thu Apr 28 06:45:36 PM EDT 20226. Ryzen 7 5800X: Build Time Mon Apr 25 06:10:54 PM EDT 2022

OpenBenchmarking.orgSeconds, Fewer Is BetterOpen Porous Media GitOPM Benchmark: Flow MPI Norne - Threads: 8Ryzen 9 5950XRyzen 9 5900XRyzen 7 5800X3DRyzen 7 5800XCore i9 12900K110220330440550SE +/- 0.16, N = 3SE +/- 0.23, N = 3SE +/- 0.25, N = 3SE +/- 0.04, N = 3SE +/- 0.26, N = 3357.86361.84361.91447.63515.711. (CXX) g++ options: -pipe -pthread -fopenmp -O3 -mtune=native -UNDEBUG -lm -ldl -lrt2. Ryzen 9 5950X: Build Time Mon Apr 25 06:10:54 PM EDT 20223. Ryzen 9 5900X: Build Time Mon Apr 25 06:10:54 PM EDT 20224. Ryzen 7 5800X3D: Build Time Mon Apr 25 06:10:54 PM EDT 20225. Ryzen 7 5800X: Build Time Mon Apr 25 06:10:54 PM EDT 20226. Core i9 12900K: Build Time Thu Apr 28 06:45:36 PM EDT 2022

OpenBenchmarking.orgSeconds, Fewer Is BetterOpen Porous Media GitOPM Benchmark: Flow MPI Norne-4C MSW - Threads: 4Ryzen 7 5800X3DRyzen 9 5950XRyzen 9 5900XCore i9 12900KRyzen 7 5800X100200300400500SE +/- 0.21, N = 3SE +/- 0.39, N = 3SE +/- 0.78, N = 3SE +/- 0.53, N = 3SE +/- 0.49, N = 3357.33377.03378.33451.00454.341. (CXX) g++ options: -pipe -pthread -fopenmp -O3 -mtune=native -UNDEBUG -lm -ldl -lrt2. Ryzen 7 5800X3D: Build Time Mon Apr 25 06:10:54 PM EDT 20223. Ryzen 9 5950X: Build Time Mon Apr 25 06:10:54 PM EDT 20224. Ryzen 9 5900X: Build Time Mon Apr 25 06:10:54 PM EDT 20225. Core i9 12900K: Build Time Thu Apr 28 06:45:36 PM EDT 20226. Ryzen 7 5800X: Build Time Mon Apr 25 06:10:54 PM EDT 2022

ONNX Runtime

ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: yolov4 - Device: CPU - Executor: StandardCore i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X150300450600750SE +/- 1.42, N = 3SE +/- 42.62, N = 12SE +/- 17.59, N = 12SE +/- 17.40, N = 9SE +/- 28.87, N = 126935725394874311. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

LeelaChessZero

LeelaChessZero (lc0 / lczero) is a chess engine automated vian neural networks. This test profile can be used for OpenCL, CUDA + cuDNN, and BLAS (CPU-based) benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgNodes Per Second, More Is BetterLeelaChessZero 0.28Backend: BLASCore i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 7 5800XRyzen 9 5950X5001000150020002500SE +/- 22.18, N = 3SE +/- 6.08, N = 3SE +/- 11.02, N = 3SE +/- 6.36, N = 3SE +/- 2.73, N = 3220112549548676681. (CXX) g++ options: -flto -pthread

ECP-CANDLE

The CANDLE benchmark codes implement deep learning architectures relevant to problems in cancer. These architectures address problems at different biological scales, specifically problems at the molecular, cellular and population scales. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterECP-CANDLE 0.4Benchmark: P3B1Core i9 12900KRyzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XRyzen 9 5950X30060090012001500429.611023.321158.671174.471309.59

WebP2 Image Encode

This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterWebP2 Image Encode 20220422Encode Settings: Quality 95, Compression Effort 7Core i9 12900KRyzen 9 5950XRyzen 9 5900XRyzen 7 5800X3DRyzen 7 5800X90180270360450SE +/- 0.51, N = 3SE +/- 1.17, N = 3SE +/- 0.76, N = 3SE +/- 0.77, N = 3SE +/- 1.68, N = 3215.58234.93268.61376.96420.031. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl

Open Porous Media Git

This is a test of Open Porous Media, a set of open-source tools concerning simulation of flow and transport of fluids in porous media. This test profile builds OPM and its dependencies from upstream Git. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpen Porous Media GitOPM Benchmark: Flow MPI Norne - Threads: 1Ryzen 7 5800X3DCore i9 12900KRyzen 9 5950XRyzen 9 5900XRyzen 7 5800X60120180240300SE +/- 0.10, N = 3SE +/- 0.07, N = 3SE +/- 1.36, N = 3SE +/- 0.23, N = 3SE +/- 0.17, N = 3224.10247.89270.81273.89281.471. (CXX) g++ options: -pipe -pthread -fopenmp -O3 -mtune=native -UNDEBUG -lm -ldl -lrt2. Ryzen 7 5800X3D: Build Time Mon Apr 25 06:10:54 PM EDT 20223. Core i9 12900K: Build Time Thu Apr 28 06:45:36 PM EDT 20224. Ryzen 9 5950X: Build Time Mon Apr 25 06:10:54 PM EDT 20225. Ryzen 9 5900X: Build Time Mon Apr 25 06:10:54 PM EDT 20226. Ryzen 7 5800X: Build Time Mon Apr 25 06:10:54 PM EDT 2022

OpenBenchmarking.orgSeconds, Fewer Is BetterOpen Porous Media GitOPM Benchmark: Flow MPI Norne - Threads: 4Ryzen 7 5800X3DRyzen 9 5950XRyzen 9 5900XRyzen 7 5800XCore i9 12900K60120180240300SE +/- 0.14, N = 3SE +/- 0.31, N = 3SE +/- 0.18, N = 3SE +/- 0.16, N = 3SE +/- 0.28, N = 3223.97232.61233.63275.40280.851. (CXX) g++ options: -pipe -pthread -fopenmp -O3 -mtune=native -UNDEBUG -lm -ldl -lrt2. Ryzen 7 5800X3D: Build Time Mon Apr 25 06:10:54 PM EDT 20223. Ryzen 9 5950X: Build Time Mon Apr 25 06:10:54 PM EDT 20224. Ryzen 9 5900X: Build Time Mon Apr 25 06:10:54 PM EDT 20225. Ryzen 7 5800X: Build Time Mon Apr 25 06:10:54 PM EDT 20226. Core i9 12900K: Build Time Thu Apr 28 06:45:36 PM EDT 2022

ECP-CANDLE

The CANDLE benchmark codes implement deep learning architectures relevant to problems in cancer. These architectures address problems at different biological scales, specifically problems at the molecular, cellular and population scales. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterECP-CANDLE 0.4Benchmark: P3B2Ryzen 7 5800XRyzen 7 5800X3DRyzen 9 5950XRyzen 9 5900XCore i9 12900K30060090012001500390.27553.75654.35667.281503.26

ONNX Runtime

ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: ArcFace ResNet-100 - Device: CPU - Executor: StandardCore i9 12900KRyzen 7 5800X3DRyzen 9 5950XRyzen 9 5900XRyzen 7 5800X400800120016002000SE +/- 0.67, N = 3SE +/- 3.42, N = 3SE +/- 46.80, N = 9SE +/- 45.67, N = 12SE +/- 1.33, N = 3197519391668151910421. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: GPT-2 - Device: CPU - Executor: StandardCore i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X2K4K6K8K10KSE +/- 20.71, N = 3SE +/- 14.95, N = 3SE +/- 10.85, N = 3SE +/- 58.87, N = 8SE +/- 104.18, N = 121108288267862706268321. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

Open Porous Media Git

This is a test of Open Porous Media, a set of open-source tools concerning simulation of flow and transport of fluids in porous media. This test profile builds OPM and its dependencies from upstream Git. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpen Porous Media GitOPM Benchmark: Flow MPI Norne - Threads: 2Ryzen 7 5800X3DRyzen 9 5950XRyzen 9 5900XCore i9 12900KRyzen 7 5800X50100150200250SE +/- 0.14, N = 3SE +/- 0.42, N = 3SE +/- 0.16, N = 3SE +/- 0.42, N = 3SE +/- 0.53, N = 3187.49203.89203.98223.20234.631. (CXX) g++ options: -pipe -pthread -fopenmp -O3 -mtune=native -UNDEBUG -lm -ldl -lrt2. Ryzen 7 5800X3D: Build Time Mon Apr 25 06:10:54 PM EDT 20223. Ryzen 9 5950X: Build Time Mon Apr 25 06:10:54 PM EDT 20224. Ryzen 9 5900X: Build Time Mon Apr 25 06:10:54 PM EDT 20225. Core i9 12900K: Build Time Thu Apr 28 06:45:36 PM EDT 20226. Ryzen 7 5800X: Build Time Mon Apr 25 06:10:54 PM EDT 2022

ONNX Runtime

ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: fcn-resnet101-11 - Device: CPU - Executor: StandardRyzen 9 5900XRyzen 7 5800X3DCore i9 12900KRyzen 9 5950XRyzen 7 5800X306090120150SE +/- 0.44, N = 3SE +/- 0.17, N = 3SE +/- 0.00, N = 3SE +/- 4.73, N = 12SE +/- 0.00, N = 311510710198491. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: bertsquad-12 - Device: CPU - Executor: StandardCore i9 12900KRyzen 9 5900XRyzen 7 5800X3DRyzen 9 5950XRyzen 7 5800X2004006008001000SE +/- 0.76, N = 3SE +/- 4.80, N = 3SE +/- 57.02, N = 12SE +/- 1.20, N = 3SE +/- 0.17, N = 39889498228005571. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: super-resolution-10 - Device: CPU - Executor: StandardRyzen 9 5900XRyzen 9 5950XCore i9 12900KRyzen 7 5800X3DRyzen 7 5800X16003200480064008000SE +/- 318.33, N = 12SE +/- 21.53, N = 3SE +/- 10.33, N = 3SE +/- 11.77, N = 3SE +/- 10.27, N = 3754561694747410736281. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210720Target: CPU - Model: regnety_400mRyzen 7 5800X3DRyzen 7 5800XCore i9 12900KRyzen 9 5900XRyzen 9 5950X3691215SE +/- 0.01, N = 15SE +/- 0.01, N = 15SE +/- 0.22, N = 15SE +/- 0.01, N = 3SE +/- 0.08, N = 45.185.937.398.439.61MIN: 5.07 / MAX: 12.17MIN: 5.83 / MAX: 7.68MIN: 6.17 / MAX: 27.59MIN: 8.36 / MAX: 8.75MIN: 9.37 / MAX: 11.031. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210720Target: CPU - Model: squeezenet_ssdRyzen 7 5800X3DCore i9 12900KRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X48121620SE +/- 0.05, N = 15SE +/- 0.17, N = 15SE +/- 0.02, N = 3SE +/- 0.09, N = 4SE +/- 0.06, N = 1512.4413.2713.5114.5516.76MIN: 12.03 / MAX: 14.14MIN: 12.19 / MAX: 43.4MIN: 13.16 / MAX: 20.71MIN: 13.65 / MAX: 21.26MIN: 16.11 / MAX: 23.011. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210720Target: CPU - Model: yolov4-tinyRyzen 7 5800X3DCore i9 12900KRyzen 7 5800XRyzen 9 5950XRyzen 9 5900X510152025SE +/- 0.19, N = 15SE +/- 0.33, N = 15SE +/- 0.21, N = 15SE +/- 0.32, N = 4SE +/- 0.40, N = 314.8015.8619.7820.4920.63MIN: 14 / MAX: 16.98MIN: 14.24 / MAX: 21MIN: 18.64 / MAX: 21.1MIN: 19.6 / MAX: 21.74MIN: 19.48 / MAX: 21.61. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210720Target: CPU - Model: resnet50Core i9 12900KRyzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XRyzen 9 5950X612182430SE +/- 0.07, N = 15SE +/- 0.09, N = 15SE +/- 0.07, N = 15SE +/- 0.04, N = 3SE +/- 0.34, N = 416.8418.1320.2221.2124.40MIN: 16.32 / MAX: 21.73MIN: 17.64 / MAX: 24.72MIN: 19.74 / MAX: 28.22MIN: 20.94 / MAX: 23.21MIN: 23.71 / MAX: 27.091. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210720Target: CPU - Model: alexnetCore i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X3691215SE +/- 0.05, N = 15SE +/- 0.03, N = 15SE +/- 0.01, N = 3SE +/- 0.13, N = 4SE +/- 0.01, N = 157.589.6210.0111.0811.83MIN: 7.18 / MAX: 9.34MIN: 8.9 / MAX: 11.14MIN: 9.92 / MAX: 12.23MIN: 10.7 / MAX: 12.6MIN: 11.63 / MAX: 18.461. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210720Target: CPU - Model: resnet18Core i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 7 5800XRyzen 9 5950X48121620SE +/- 0.15, N = 15SE +/- 0.05, N = 15SE +/- 0.05, N = 3SE +/- 0.03, N = 15SE +/- 0.17, N = 49.6610.2712.5013.0214.28MIN: 7.55 / MAX: 14.7MIN: 9.71 / MAX: 12.09MIN: 12.28 / MAX: 12.82MIN: 12.77 / MAX: 32.91MIN: 13.94 / MAX: 16.831. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210720Target: CPU - Model: vgg16Core i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 7 5800XRyzen 9 5950X1326395265SE +/- 0.48, N = 15SE +/- 0.13, N = 15SE +/- 0.09, N = 3SE +/- 0.07, N = 15SE +/- 0.08, N = 428.2442.6050.7555.9756.55MIN: 25.72 / MAX: 45.6MIN: 41.52 / MAX: 50.92MIN: 49.97 / MAX: 60.2MIN: 54.91 / MAX: 64.62MIN: 55.53 / MAX: 62.981. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210720Target: CPU - Model: googlenetRyzen 7 5800X3DCore i9 12900KRyzen 7 5800XRyzen 9 5900XRyzen 9 5950X3691215SE +/- 0.05, N = 15SE +/- 0.21, N = 15SE +/- 0.02, N = 15SE +/- 0.02, N = 3SE +/- 0.28, N = 47.319.9410.2211.4412.92MIN: 7.05 / MAX: 14.4MIN: 7.91 / MAX: 14.3MIN: 9.79 / MAX: 18.21MIN: 11.28 / MAX: 11.82MIN: 12.09 / MAX: 15.321. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210720Target: CPU - Model: blazefaceRyzen 7 5800X3DRyzen 7 5800XCore i9 12900KRyzen 9 5900XRyzen 9 5950X0.4050.811.2151.622.025SE +/- 0.00, N = 15SE +/- 0.00, N = 15SE +/- 0.05, N = 15SE +/- 0.00, N = 3SE +/- 0.02, N = 41.061.221.461.631.80MIN: 1.04 / MAX: 4.31MIN: 1.19 / MAX: 2.1MIN: 1.15 / MAX: 2.96MIN: 1.61 / MAX: 1.81MIN: 1.75 / MAX: 2.151. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210720Target: CPU - Model: efficientnet-b0Ryzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XRyzen 9 5950XCore i9 12900K1.20152.4033.60454.8066.0075SE +/- 0.01, N = 15SE +/- 0.01, N = 15SE +/- 0.01, N = 3SE +/- 0.06, N = 4SE +/- 0.11, N = 153.013.604.775.235.34MIN: 2.93 / MAX: 13.07MIN: 3.53 / MAX: 5.22MIN: 4.7 / MAX: 5MIN: 5.09 / MAX: 6.68MIN: 4.35 / MAX: 9.281. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210720Target: CPU - Model: mnasnetRyzen 7 5800X3DRyzen 7 5800XCore i9 12900KRyzen 9 5900XRyzen 9 5950X0.87081.74162.61243.48324.354SE +/- 0.00, N = 15SE +/- 0.00, N = 15SE +/- 0.07, N = 15SE +/- 0.01, N = 3SE +/- 0.05, N = 42.012.243.103.443.87MIN: 1.97 / MAX: 2.76MIN: 2.2 / MAX: 3.68MIN: 2.66 / MAX: 4.79MIN: 3.39 / MAX: 3.72MIN: 3.76 / MAX: 10.71. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210720Target: CPU - Model: shufflenet-v2Ryzen 7 5800X3DRyzen 7 5800XCore i9 12900KRyzen 9 5900XRyzen 9 5950X0.9361.8722.8083.7444.68SE +/- 0.01, N = 15SE +/- 0.00, N = 15SE +/- 0.08, N = 14SE +/- 0.01, N = 3SE +/- 0.01, N = 42.112.353.103.884.16MIN: 2.07 / MAX: 3MIN: 2.31 / MAX: 3.74MIN: 2.68 / MAX: 4.51MIN: 3.82 / MAX: 4.06MIN: 4.05 / MAX: 4.91. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210720Target: CPU-v3-v3 - Model: mobilenet-v3Ryzen 7 5800X3DRyzen 7 5800XCore i9 12900KRyzen 9 5900XRyzen 9 5950X0.84831.69662.54493.39324.2415SE +/- 0.00, N = 15SE +/- 0.00, N = 15SE +/- 0.06, N = 15SE +/- 0.01, N = 3SE +/- 0.00, N = 41.902.282.903.463.77MIN: 1.84 / MAX: 2.38MIN: 2.22 / MAX: 3.92MIN: 2.53 / MAX: 4.55MIN: 3.39 / MAX: 3.66MIN: 3.7 / MAX: 4.731. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210720Target: CPU-v2-v2 - Model: mobilenet-v2Ryzen 7 5800X3DRyzen 7 5800XCore i9 12900KRyzen 9 5900XRyzen 9 5950X0.96531.93062.89593.86124.8265SE +/- 0.00, N = 15SE +/- 0.01, N = 15SE +/- 0.12, N = 15SE +/- 0.01, N = 3SE +/- 0.01, N = 42.132.613.413.914.29MIN: 2.03 / MAX: 2.88MIN: 2.54 / MAX: 3.77MIN: 2.72 / MAX: 5.86MIN: 3.82 / MAX: 4.11MIN: 4.15 / MAX: 7.551. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20210720Target: CPU - Model: mobilenetRyzen 7 5800X3DCore i9 12900KRyzen 9 5900XRyzen 7 5800XRyzen 9 5950X3691215SE +/- 0.08, N = 15SE +/- 0.26, N = 15SE +/- 0.01, N = 3SE +/- 0.13, N = 15SE +/- 0.15, N = 47.6211.1111.4211.6312.27MIN: 7.25 / MAX: 9.92MIN: 8.87 / MAX: 13.96MIN: 11.16 / MAX: 18.48MIN: 11.06 / MAX: 12.93MIN: 11.71 / MAX: 18.811. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

Xcompact3d Incompact3d

Xcompact3d Incompact3d is a Fortran-MPI based, finite difference high-performance code for solving the incompressible Navier-Stokes equation and as many as you need scalar transport equations. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterXcompact3d Incompact3d 2021-03-11Input: input.i3d 193 Cells Per DirectionCore i9 12900KRyzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XRyzen 9 5950X306090120150SE +/- 0.58, N = 3SE +/- 0.11, N = 3SE +/- 1.17, N = 9SE +/- 0.05, N = 3SE +/- 0.16, N = 355.61126.86141.21144.47156.161. (F9X) gfortran options: -cpp -O2 -funroll-loops -floop-optimize -fcray-pointer -fbacktrace -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz

TNN

TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: DenseNetCore i9 12900KRyzen 9 5950XRyzen 9 5900XRyzen 7 5800X3DRyzen 7 5800X6001200180024003000SE +/- 1.88, N = 3SE +/- 10.08, N = 3SE +/- 3.11, N = 3SE +/- 1.54, N = 3SE +/- 0.87, N = 31792.972425.592563.762609.843016.48MIN: 1751.55 / MAX: 1868.31MIN: 2339.59 / MAX: 2518.62MIN: 2481.34 / MAX: 2640.2MIN: 2559.5 / MAX: 2657.73MIN: 2943.1 / MAX: 3087.971. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl

Caffe

This is a benchmark of the Caffe deep learning framework and currently supports the AlexNet and Googlenet model and execution on both CPUs and NVIDIA GPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMilli-Seconds, Fewer Is BetterCaffe 2020-02-13Model: GoogleNet - Acceleration: CPU - Iterations: 200Core i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 7 5800XRyzen 9 5950X40K80K120K160K200KSE +/- 116.45, N = 3SE +/- 248.50, N = 3SE +/- 46.44, N = 3SE +/- 39.00, N = 3SE +/- 215.28, N = 31340451620221787251791461937101. (CXX) g++ options: -fPIC -O3 -rdynamic -lglog -lgflags -lprotobuf -lcrypto -lcurl -lpthread -lsz -lz -ldl -lm -llmdb -lopenblas

Numpy Benchmark

This is a test to obtain the general Numpy performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgScore, More Is BetterNumpy BenchmarkCore i9 12900KRyzen 9 5900XRyzen 7 5800X3DRyzen 9 5950XRyzen 7 5800X140280420560700SE +/- 0.68, N = 3SE +/- 8.26, N = 3SE +/- 1.21, N = 3SE +/- 0.86, N = 3SE +/- 0.43, N = 3665.34605.17603.09591.54491.56

WebP2 Image Encode

This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterWebP2 Image Encode 20220422Encode Settings: Quality 75, Compression Effort 7Core i9 12900KRyzen 9 5950XRyzen 9 5900XRyzen 7 5800X3DRyzen 7 5800X4080120160200SE +/- 0.23, N = 3SE +/- 1.48, N = 3SE +/- 0.51, N = 3SE +/- 0.62, N = 3SE +/- 0.74, N = 3101.57110.28129.73178.01199.651. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl

ONNX Runtime

ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: super-resolution-10 - Device: CPU - Executor: ParallelRyzen 9 5950XRyzen 9 5900XRyzen 7 5800X3DCore i9 12900KRyzen 7 5800X14002800420056007000SE +/- 22.98, N = 3SE +/- 5.61, N = 3SE +/- 15.67, N = 3SE +/- 46.82, N = 4SE +/- 10.11, N = 3643655914606451641031. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: fcn-resnet101-11 - Device: CPU - Executor: ParallelCore i9 12900KRyzen 9 5950XRyzen 9 5900XRyzen 7 5800X3DRyzen 7 5800X20406080100SE +/- 0.44, N = 3SE +/- 0.33, N = 3SE +/- 0.00, N = 3SE +/- 0.17, N = 3SE +/- 0.29, N = 3111888572671. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: GPT-2 - Device: CPU - Executor: ParallelCore i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 7 5800XRyzen 9 5950X2K4K6K8K10KSE +/- 20.34, N = 3SE +/- 5.78, N = 3SE +/- 11.00, N = 3SE +/- 8.85, N = 3SE +/- 11.18, N = 3808869195687562155341. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: ArcFace ResNet-100 - Device: CPU - Executor: ParallelRyzen 9 5950XRyzen 9 5900XRyzen 7 5800X3DRyzen 7 5800XCore i9 12900K30060090012001500SE +/- 8.09, N = 3SE +/- 1.17, N = 3SE +/- 3.51, N = 3SE +/- 4.25, N = 3SE +/- 0.17, N = 315401421122311073621. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: yolov4 - Device: CPU - Executor: ParallelCore i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X140280420560700SE +/- 1.09, N = 3SE +/- 0.60, N = 3SE +/- 0.33, N = 3SE +/- 0.44, N = 3SE +/- 0.44, N = 36293082992962731. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: bertsquad-12 - Device: CPU - Executor: ParallelCore i9 12900KRyzen 9 5950XRyzen 9 5900XRyzen 7 5800X3DRyzen 7 5800X2004006008001000SE +/- 11.29, N = 3SE +/- 0.29, N = 3SE +/- 0.60, N = 3SE +/- 1.17, N = 3SE +/- 0.29, N = 39195545405184801. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

ASKAP

ASKAP is a set of benchmarks from the Australian SKA Pathfinder. The principal ASKAP benchmarks are the Hogbom Clean Benchmark (tHogbomClean) and Convolutional Resamping Benchmark (tConvolve) as well as some previous ASKAP benchmarks being included as well for OpenCL and CUDA execution of tConvolve. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 1.0Test: tConvolve MT - DegriddingCore i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 7 5800XRyzen 9 5950X8001600240032004000SE +/- 2.07, N = 3SE +/- 1.83, N = 3SE +/- 3.45, N = 3SE +/- 5.57, N = 3SE +/- 0.92, N = 33872.031594.951541.671471.581346.571. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 1.0Test: tConvolve MT - GriddingCore i9 12900KRyzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XRyzen 9 5950X6001200180024003000SE +/- 2.70, N = 3SE +/- 2.64, N = 3SE +/- 2.02, N = 3SE +/- 0.90, N = 3SE +/- 1.65, N = 32721.15930.57854.71837.94784.511. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp

OpenFOAM

OpenFOAM is the leading free, open source software for computational fluid dynamics (CFD). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenFOAM 8Input: Motorbike 30MRyzen 7 5800X3DCore i9 12900KRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X4080120160200SE +/- 0.32, N = 3SE +/- 0.18, N = 3SE +/- 0.15, N = 3SE +/- 0.24, N = 3SE +/- 0.25, N = 380.2984.4796.1698.44177.60-lfoamToVTK -llagrangian -lfileFormats1. (CXX) g++ options: -std=c++11 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -ldynamicMesh -lgenericPatchFields -lOpenFOAM -ldl -lm

Caffe

This is a benchmark of the Caffe deep learning framework and currently supports the AlexNet and Googlenet model and execution on both CPUs and NVIDIA GPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMilli-Seconds, Fewer Is BetterCaffe 2020-02-13Model: GoogleNet - Acceleration: CPU - Iterations: 100Core i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 7 5800XRyzen 9 5950X20K40K60K80K100KSE +/- 741.46, N = 4SE +/- 158.22, N = 3SE +/- 260.55, N = 3SE +/- 34.64, N = 3SE +/- 34.42, N = 367833808628949289825966241. (CXX) g++ options: -fPIC -O3 -rdynamic -lglog -lgflags -lprotobuf -lcrypto -lcurl -lpthread -lsz -lz -ldl -lm -llmdb -lopenblas

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPURyzen 7 5800X3DCore i9 12900KRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X400800120016002000SE +/- 1.81, N = 3SE +/- 2.95, N = 3SE +/- 5.06, N = 3SE +/- 19.76, N = 5SE +/- 9.79, N = 31385.791617.101776.191820.401847.70-lpthread - MIN: 1375.42MIN: 1608.6-lpthread - MIN: 1756.45-lpthread - MIN: 1761.07-lpthread - MIN: 1820.321. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPURyzen 7 5800X3DRyzen 9 5950XCore i9 12900KRyzen 9 5900XRyzen 7 5800X7001400210028003500SE +/- 2.34, N = 3SE +/- 26.41, N = 3SE +/- 0.76, N = 3SE +/- 32.41, N = 3SE +/- 1.56, N = 32691.242750.182881.322904.333065.26-lpthread - MIN: 2678.39-lpthread - MIN: 2685.91MIN: 2872.06-lpthread - MIN: 2837.45-lpthread - MIN: 3058.731. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPURyzen 7 5800X3DRyzen 9 5950XRyzen 9 5900XCore i9 12900KRyzen 7 5800X7001400210028003500SE +/- 5.12, N = 3SE +/- 10.92, N = 3SE +/- 23.67, N = 3SE +/- 0.27, N = 3SE +/- 2.27, N = 32683.162703.332873.452881.723053.30-lpthread - MIN: 2665.66-lpthread - MIN: 2665.5-lpthread - MIN: 2830.55MIN: 2874.65-lpthread - MIN: 3045.761. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPURyzen 7 5800X3DRyzen 9 5950XCore i9 12900KRyzen 9 5900XRyzen 7 5800X7001400210028003500SE +/- 1.77, N = 3SE +/- 23.91, N = 3SE +/- 1.61, N = 3SE +/- 15.10, N = 3SE +/- 0.89, N = 32691.672745.442881.102908.283055.84-lpthread - MIN: 2680.17-lpthread - MIN: 2684.63MIN: 2869.73-lpthread - MIN: 2862.45-lpthread - MIN: 3050.911. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

Mlpack Benchmark

Mlpack benchmark scripts for machine learning libraries Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterMlpack BenchmarkBenchmark: scikit_qdaCore i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 7 5800XRyzen 9 5950X1224364860SE +/- 0.13, N = 3SE +/- 0.08, N = 3SE +/- 0.34, N = 3SE +/- 0.10, N = 3SE +/- 0.20, N = 327.3138.2751.3952.5755.06

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPURyzen 7 5800X3DCore i9 12900KRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X400800120016002000SE +/- 2.31, N = 3SE +/- 0.28, N = 3SE +/- 8.68, N = 3SE +/- 8.84, N = 3SE +/- 5.60, N = 31382.161613.811785.051814.131859.14-lpthread - MIN: 1372.35MIN: 1608.23-lpthread - MIN: 1762.05-lpthread - MIN: 1783.13-lpthread - MIN: 1846.971. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPURyzen 7 5800X3DCore i9 12900KRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X400800120016002000SE +/- 0.61, N = 3SE +/- 1.95, N = 3SE +/- 13.72, N = 3SE +/- 22.20, N = 3SE +/- 4.89, N = 31387.861616.061742.061783.851864.54-lpthread - MIN: 1380.9MIN: 1608.68-lpthread - MIN: 1715.13-lpthread - MIN: 1730.18-lpthread - MIN: 1851.661. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: inception-v3Ryzen 7 5800X3DRyzen 9 5900XCore i9 12900KRyzen 7 5800XRyzen 9 5950X612182430SE +/- 0.09, N = 3SE +/- 0.37, N = 3SE +/- 0.66, N = 3SE +/- 0.09, N = 3SE +/- 0.13, N = 323.1923.6924.3025.6925.73MIN: 22.93 / MAX: 31MIN: 22.99 / MAX: 31.36MIN: 22.9 / MAX: 36.09MIN: 25.45 / MAX: 31.44MIN: 25.09 / MAX: 33.941. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: mobilenet-v1-1.0Ryzen 7 5800X3DRyzen 7 5800XRyzen 9 5950XCore i9 12900KRyzen 9 5900X0.91621.83242.74863.66484.581SE +/- 0.010, N = 3SE +/- 0.011, N = 3SE +/- 0.067, N = 3SE +/- 0.008, N = 3SE +/- 0.037, N = 31.6761.8162.6392.8914.072MIN: 1.63 / MAX: 2.97MIN: 1.79 / MAX: 3.08MIN: 2.52 / MAX: 11.27MIN: 2.85 / MAX: 8.54MIN: 3.95 / MAX: 4.311. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: MobileNetV2_224Ryzen 7 5800X3DRyzen 7 5800XCore i9 12900KRyzen 9 5900XRyzen 9 5950X0.73691.47382.21072.94763.6845SE +/- 0.011, N = 3SE +/- 0.029, N = 3SE +/- 0.020, N = 3SE +/- 0.045, N = 3SE +/- 0.014, N = 31.8311.9422.4102.9753.275MIN: 1.79 / MAX: 2.92MIN: 1.9 / MAX: 3.6MIN: 2.36 / MAX: 3.86MIN: 2.88 / MAX: 3.44MIN: 3.21 / MAX: 10.861. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: SqueezeNetV1.0Core i9 12900KRyzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XRyzen 9 5950X1.1772.3543.5314.7085.885SE +/- 0.177, N = 3SE +/- 0.010, N = 3SE +/- 0.032, N = 3SE +/- 0.058, N = 3SE +/- 0.113, N = 34.1514.2134.5364.7585.231MIN: 3.91 / MAX: 6.32MIN: 4.16 / MAX: 5.46MIN: 4.47 / MAX: 5.71MIN: 4.59 / MAX: 12.51MIN: 4.95 / MAX: 6.511. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: resnet-v2-50Ryzen 7 5800X3DRyzen 7 5800XRyzen 9 5950XCore i9 12900KRyzen 9 5900X612182430SE +/- 0.10, N = 3SE +/- 0.03, N = 3SE +/- 0.14, N = 3SE +/- 1.14, N = 3SE +/- 0.14, N = 316.3618.4420.5223.0924.50MIN: 16 / MAX: 24.17MIN: 18.25 / MAX: 25.86MIN: 19.84 / MAX: 24.05MIN: 21.7 / MAX: 30.17MIN: 23.86 / MAX: 52.441. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: squeezenetv1.1Core i9 12900KRyzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XRyzen 9 5950X0.89731.79462.69193.58924.4865SE +/- 0.049, N = 3SE +/- 0.006, N = 3SE +/- 0.017, N = 3SE +/- 0.098, N = 3SE +/- 0.098, N = 32.4052.5902.8053.3943.988MIN: 2.33 / MAX: 3.56MIN: 2.55 / MAX: 4.48MIN: 2.76 / MAX: 10.42MIN: 3.15 / MAX: 4.21MIN: 3.72 / MAX: 4.791. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 1.2Model: mobilenetV3Ryzen 7 5800X3DRyzen 7 5800XCore i9 12900KRyzen 9 5900XRyzen 9 5950X0.5381.0761.6142.1522.69SE +/- 0.006, N = 3SE +/- 0.005, N = 3SE +/- 0.007, N = 3SE +/- 0.018, N = 3SE +/- 0.003, N = 31.0821.1561.1741.8382.391MIN: 1.06 / MAX: 2.25MIN: 1.14 / MAX: 1.73MIN: 1.15 / MAX: 2.06MIN: 1.79 / MAX: 2.07MIN: 1.87 / MAX: 3.851. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPURyzen 9 5950XRyzen 9 5900XRyzen 7 5800X3DRyzen 7 5800XCore i9 12900K246810SE +/- 0.22726, N = 15SE +/- 0.07639, N = 3SE +/- 0.06264, N = 3SE +/- 0.11645, N = 15SE +/- 0.16860, N = 124.755335.386897.276238.342038.73750-lpthread - MIN: 3.38-lpthread - MIN: 4-lpthread - MIN: 5.11-lpthread - MIN: 5.88MIN: 4.151. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

Mlpack Benchmark

Mlpack benchmark scripts for machine learning libraries Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterMlpack BenchmarkBenchmark: scikit_linearridgeregressionRyzen 9 5900XCore i9 12900KRyzen 7 5800X3DRyzen 9 5950XRyzen 7 5800X0.40950.8191.22851.6382.0475SE +/- 0.00, N = 3SE +/- 0.05, N = 12SE +/- 0.00, N = 3SE +/- 0.01, N = 3SE +/- 0.02, N = 71.581.601.631.691.82

Caffe

This is a benchmark of the Caffe deep learning framework and currently supports the AlexNet and Googlenet model and execution on both CPUs and NVIDIA GPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMilli-Seconds, Fewer Is BetterCaffe 2020-02-13Model: AlexNet - Acceleration: CPU - Iterations: 200Core i9 12900KRyzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XRyzen 9 5950X16K32K48K64K80KSE +/- 323.04, N = 3SE +/- 94.00, N = 3SE +/- 49.12, N = 3SE +/- 57.49, N = 3SE +/- 178.17, N = 351713604946766168676733521. (CXX) g++ options: -fPIC -O3 -rdynamic -lglog -lgflags -lprotobuf -lcrypto -lcurl -lpthread -lsz -lz -ldl -lm -llmdb -lopenblas

ASKAP

ASKAP is a set of benchmarks from the Australian SKA Pathfinder. The principal ASKAP benchmarks are the Hogbom Clean Benchmark (tHogbomClean) and Convolutional Resamping Benchmark (tConvolve) as well as some previous ASKAP benchmarks being included as well for OpenCL and CUDA execution of tConvolve. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMpix/sec, More Is BetterASKAP 1.0Test: tConvolve MPI - GriddingRyzen 7 5800X3DRyzen 9 5900XRyzen 9 5950XCore i9 12900KRyzen 7 5800X2K4K6K8K10KSE +/- 0.00, N = 3SE +/- 58.16, N = 3SE +/- 12.67, N = 3SE +/- 45.52, N = 38746.528258.026728.094472.734256.051. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp

OpenBenchmarking.orgMpix/sec, More Is BetterASKAP 1.0Test: tConvolve MPI - DegriddingRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X3DCore i9 12900KRyzen 7 5800X16003200480064008000SE +/- 49.47, N = 3SE +/- 48.56, N = 3SE +/- 53.33, N = 3SE +/- 19.39, N = 3SE +/- 34.79, N = 37668.056643.646453.224198.513976.301. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp

Mlpack Benchmark

Mlpack benchmark scripts for machine learning libraries Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterMlpack BenchmarkBenchmark: scikit_icaCore i9 12900KRyzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XRyzen 9 5950X1020304050SE +/- 0.02, N = 3SE +/- 0.03, N = 3SE +/- 0.05, N = 3SE +/- 0.24, N = 3SE +/- 0.12, N = 332.9334.8839.1639.3741.77

Caffe

This is a benchmark of the Caffe deep learning framework and currently supports the AlexNet and Googlenet model and execution on both CPUs and NVIDIA GPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMilli-Seconds, Fewer Is BetterCaffe 2020-02-13Model: AlexNet - Acceleration: CPU - Iterations: 100Core i9 12900KRyzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XRyzen 9 5950X8K16K24K32K40KSE +/- 17.32, N = 3SE +/- 24.84, N = 3SE +/- 7.06, N = 3SE +/- 38.11, N = 3SE +/- 30.55, N = 325590302223390534329366581. (CXX) g++ options: -fPIC -O3 -rdynamic -lglog -lgflags -lprotobuf -lcrypto -lcurl -lpthread -lsz -lz -ldl -lm -llmdb -lopenblas

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPURyzen 9 5950XRyzen 9 5900XCore i9 12900KRyzen 7 5800X3DRyzen 7 5800X0.31660.63320.94981.26641.583SE +/- 0.006926, N = 15SE +/- 0.007658, N = 4SE +/- 0.008609, N = 4SE +/- 0.001247, N = 4SE +/- 0.000645, N = 40.8230091.0069411.0520001.2378501.407280-lpthread - MIN: 0.7-lpthread - MIN: 0.94MIN: 1.01-lpthread - MIN: 1.21-lpthread - MIN: 1.391. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

Xcompact3d Incompact3d

Xcompact3d Incompact3d is a Fortran-MPI based, finite difference high-performance code for solving the incompressible Navier-Stokes equation and as many as you need scalar transport equations. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterXcompact3d Incompact3d 2021-03-11Input: input.i3d 129 Cells Per DirectionCore i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X918273645SE +/- 0.01, N = 4SE +/- 0.01, N = 3SE +/- 0.32, N = 3SE +/- 0.05, N = 3SE +/- 0.02, N = 314.5527.1832.1533.9537.211. (F9X) gfortran options: -cpp -O2 -funroll-loops -floop-optimize -fcray-pointer -fbacktrace -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz

ASKAP

ASKAP is a set of benchmarks from the Australian SKA Pathfinder. The principal ASKAP benchmarks are the Hogbom Clean Benchmark (tHogbomClean) and Convolutional Resamping Benchmark (tConvolve) as well as some previous ASKAP benchmarks being included as well for OpenCL and CUDA execution of tConvolve. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 1.0Test: tConvolve OpenMP - DegriddingRyzen 7 5800X3DCore i9 12900KRyzen 9 5900XRyzen 7 5800XRyzen 9 5950X2K4K6K8K10KSE +/- 38.17, N = 15SE +/- 37.29, N = 6SE +/- 10.98, N = 6SE +/- 6.53, N = 15SE +/- 11.91, N = 68741.607793.773732.723367.673214.581. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 1.0Test: tConvolve OpenMP - GriddingRyzen 7 5800X3DCore i9 12900KRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X15003000450060007500SE +/- 78.12, N = 15SE +/- 34.51, N = 6SE +/- 10.74, N = 6SE +/- 27.21, N = 6SE +/- 14.53, N = 156946.114631.822936.852755.721709.251. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPURyzen 9 5950XRyzen 9 5900XCore i9 12900KRyzen 7 5800X3DRyzen 7 5800X0.45810.91621.37431.83242.2905SE +/- 0.00299, N = 3SE +/- 0.00489, N = 3SE +/- 0.00773, N = 3SE +/- 0.00274, N = 3SE +/- 0.00040, N = 31.069261.300491.352121.776412.03600-lpthread - MIN: 0.96-lpthread - MIN: 1.2MIN: 1.28-lpthread - MIN: 1.74-lpthread - MIN: 2.011. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUCore i9 12900KRyzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XRyzen 9 5950X0.87671.75342.63013.50684.3835SE +/- 0.00329, N = 4SE +/- 0.00793, N = 4SE +/- 0.00733, N = 4SE +/- 0.02871, N = 4SE +/- 0.01755, N = 42.636112.897173.291363.445483.89647MIN: 2.5-lpthread - MIN: 2.81-lpthread - MIN: 3.12-lpthread - MIN: 3.05-lpthread - MIN: 3.661. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

TNN

TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: MobileNet v2Core i9 12900KRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X3DRyzen 7 5800X60120180240300SE +/- 0.31, N = 4SE +/- 0.47, N = 4SE +/- 0.71, N = 4SE +/- 0.14, N = 4SE +/- 0.18, N = 3170.58224.20224.26233.18272.04MIN: 157.87 / MAX: 209.34MIN: 218.68 / MAX: 249.25MIN: 219.36 / MAX: 242.57MIN: 232.19 / MAX: 237.22MIN: 270.94 / MAX: 276.31. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl

Mlpack Benchmark

Mlpack benchmark scripts for machine learning libraries Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterMlpack BenchmarkBenchmark: scikit_svmCore i9 12900KRyzen 9 5900XRyzen 7 5800X3DRyzen 9 5950XRyzen 7 5800X510152025SE +/- 0.01, N = 4SE +/- 0.08, N = 3SE +/- 0.01, N = 3SE +/- 0.03, N = 3SE +/- 0.01, N = 310.5416.1516.1716.5620.29

TNN

TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: SqueezeNet v1.1Core i9 12900KRyzen 9 5950XRyzen 9 5900XRyzen 7 5800X3DRyzen 7 5800X60120180240300SE +/- 0.06, N = 5SE +/- 1.16, N = 4SE +/- 1.66, N = 4SE +/- 0.03, N = 4SE +/- 0.18, N = 3133.97213.32213.78222.33266.19MIN: 133.46 / MAX: 134.81MIN: 209.34 / MAX: 215.21MIN: 210.63 / MAX: 219.97MIN: 222.15 / MAX: 222.66MIN: 265.87 / MAX: 266.681. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl

ASKAP

ASKAP is a set of benchmarks from the Australian SKA Pathfinder. The principal ASKAP benchmarks are the Hogbom Clean Benchmark (tHogbomClean) and Convolutional Resamping Benchmark (tConvolve) as well as some previous ASKAP benchmarks being included as well for OpenCL and CUDA execution of tConvolve. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations Per Second, More Is BetterASKAP 1.0Test: Hogbom Clean OpenMPCore i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 7 5800XRyzen 9 5950X120240360480600SE +/- 0.00, N = 5SE +/- 1.80, N = 4SE +/- 0.52, N = 4SE +/- 0.57, N = 3SE +/- 1.00, N = 4540.54527.72238.67221.73220.041. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUCore i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X510152025SE +/- 0.00220, N = 7SE +/- 0.01917, N = 7SE +/- 0.29208, N = 15SE +/- 0.00896, N = 7SE +/- 0.12899, N = 75.8753612.6227016.1398016.7051018.75350MIN: 5.78-lpthread - MIN: 12.28-lpthread - MIN: 15.35-lpthread - MIN: 16.31-lpthread - MIN: 18.341. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUCore i9 12900KRyzen 7 5800X3DRyzen 7 5800XRyzen 9 5900XRyzen 9 5950X48121620SE +/- 0.00249, N = 7SE +/- 0.04088, N = 7SE +/- 0.22593, N = 15SE +/- 0.04249, N = 7SE +/- 0.02368, N = 76.0021410.3800016.3534016.7382018.03210MIN: 5.9-lpthread - MIN: 9.78-lpthread - MIN: 14.78-lpthread - MIN: 16.08-lpthread - MIN: 17.61. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUCore i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X246810SE +/- 0.00441, N = 5SE +/- 0.01065, N = 5SE +/- 0.02525, N = 5SE +/- 0.05069, N = 5SE +/- 0.01310, N = 53.446436.557657.548437.796528.93552MIN: 3.4-lpthread - MIN: 6.32-lpthread - MIN: 7.29-lpthread - MIN: 7.29-lpthread - MIN: 8.331. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPURyzen 9 5950XRyzen 9 5900XRyzen 7 5800X3DCore i9 12900KRyzen 7 5800X0.32790.65580.98371.31161.6395SE +/- 0.002527, N = 5SE +/- 0.003942, N = 5SE +/- 0.002108, N = 5SE +/- 0.002976, N = 5SE +/- 0.002722, N = 50.4760970.5197330.6047800.8120471.457520-lpthread - MIN: 0.42-lpthread - MIN: 0.47-lpthread - MIN: 0.58MIN: 0.79-lpthread - MIN: 1.351. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

WebP2 Image Encode

This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterWebP2 Image Encode 20220422Encode Settings: Quality 100, Compression Effort 5Core i9 12900KRyzen 9 5950XRyzen 9 5900XRyzen 7 5800X3DRyzen 7 5800X1.32192.64383.96575.28766.6095SE +/- 0.003, N = 9SE +/- 0.006, N = 9SE +/- 0.003, N = 8SE +/- 0.003, N = 7SE +/- 0.004, N = 72.9363.2493.7095.1695.8751. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl

TNN

TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: SqueezeNet v2Core i9 12900KRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X3DRyzen 7 5800X1428425670SE +/- 0.08, N = 10SE +/- 0.22, N = 9SE +/- 0.08, N = 9SE +/- 0.13, N = 9SE +/- 0.12, N = 838.9250.8550.9453.2263.61MIN: 38.37 / MAX: 39.92MIN: 49.98 / MAX: 52.59MIN: 50.34 / MAX: 52.46MIN: 52.43 / MAX: 54.16MIN: 62.79 / MAX: 64.281. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl

ECP-CANDLE

The CANDLE benchmark codes implement deep learning architectures relevant to problems in cancer. These architectures address problems at different biological scales, specifically problems at the molecular, cellular and population scales. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterECP-CANDLE 0.4Benchmark: P1B2Core i9 12900KRyzen 7 5800X3DRyzen 9 5900XRyzen 9 5950XRyzen 7 5800X81624324020.9429.9930.2431.3934.84

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPURyzen 9 5950XRyzen 9 5900XCore i9 12900KRyzen 7 5800X3DRyzen 7 5800X246810SE +/- 0.00224, N = 9SE +/- 0.00484, N = 9SE +/- 0.00117, N = 9SE +/- 0.00483, N = 9SE +/- 0.00140, N = 93.626094.387735.253945.571096.37023-lpthread - MIN: 3.43-lpthread - MIN: 4.17MIN: 5.16-lpthread - MIN: 5.45-lpthread - MIN: 6.321. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPURyzen 9 5950XRyzen 9 5900XCore i9 12900KRyzen 7 5800X3DRyzen 7 5800X0.63881.27761.91642.55523.194SE +/- 0.00702, N = 9SE +/- 0.00298, N = 9SE +/- 0.00121, N = 9SE +/- 0.00503, N = 9SE +/- 0.00474, N = 91.618991.997422.223362.483932.83913-lpthread - MIN: 1.45-lpthread - MIN: 1.82MIN: 2.2-lpthread - MIN: 2.41-lpthread - MIN: 2.81. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

WebP2 Image Encode

This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterWebP2 Image Encode 20220422Encode Settings: DefaultCore i9 12900KRyzen 9 5950XRyzen 9 5900XRyzen 7 5800X3DRyzen 7 5800X0.81431.62862.44293.25724.0715SE +/- 0.010, N = 11SE +/- 0.013, N = 10SE +/- 0.008, N = 10SE +/- 0.007, N = 9SE +/- 0.006, N = 82.0622.1652.4393.1703.6191. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl

96 Results Shown

OpenFOAM
Open Porous Media Git
LeelaChessZero
Open Porous Media Git:
  Flow MPI Extra - 8
  Flow MPI Extra - 2
  Flow MPI Extra - 4
WebP2 Image Encode
Open Porous Media Git:
  Flow MPI Norne-4C MSW - 8
  Flow MPI Norne-4C MSW - 1
  Flow MPI Norne-4C MSW - 2
  Flow MPI Norne - 8
  Flow MPI Norne-4C MSW - 4
ONNX Runtime
LeelaChessZero
ECP-CANDLE
WebP2 Image Encode
Open Porous Media Git:
  Flow MPI Norne - 1
  Flow MPI Norne - 4
ECP-CANDLE
ONNX Runtime:
  ArcFace ResNet-100 - CPU - Standard
  GPT-2 - CPU - Standard
Open Porous Media Git
ONNX Runtime:
  fcn-resnet101-11 - CPU - Standard
  bertsquad-12 - CPU - Standard
  super-resolution-10 - CPU - Standard
NCNN:
  CPU - regnety_400m
  CPU - squeezenet_ssd
  CPU - yolov4-tiny
  CPU - resnet50
  CPU - alexnet
  CPU - resnet18
  CPU - vgg16
  CPU - googlenet
  CPU - blazeface
  CPU - efficientnet-b0
  CPU - mnasnet
  CPU - shufflenet-v2
  CPU-v3-v3 - mobilenet-v3
  CPU-v2-v2 - mobilenet-v2
  CPU - mobilenet
Xcompact3d Incompact3d
TNN
Caffe
Numpy Benchmark
WebP2 Image Encode
ONNX Runtime:
  super-resolution-10 - CPU - Parallel
  fcn-resnet101-11 - CPU - Parallel
  GPT-2 - CPU - Parallel
  ArcFace ResNet-100 - CPU - Parallel
  yolov4 - CPU - Parallel
  bertsquad-12 - CPU - Parallel
ASKAP:
  tConvolve MT - Degridding
  tConvolve MT - Gridding
OpenFOAM
Caffe
oneDNN:
  Recurrent Neural Network Inference - bf16bf16bf16 - CPU
  Recurrent Neural Network Training - bf16bf16bf16 - CPU
  Recurrent Neural Network Training - f32 - CPU
  Recurrent Neural Network Training - u8s8f32 - CPU
Mlpack Benchmark
oneDNN:
  Recurrent Neural Network Inference - f32 - CPU
  Recurrent Neural Network Inference - u8s8f32 - CPU
Mobile Neural Network:
  inception-v3
  mobilenet-v1-1.0
  MobileNetV2_224
  SqueezeNetV1.0
  resnet-v2-50
  squeezenetv1.1
  mobilenetV3
oneDNN
Mlpack Benchmark
Caffe
ASKAP:
  tConvolve MPI - Gridding
  tConvolve MPI - Degridding
Mlpack Benchmark
Caffe
oneDNN
Xcompact3d Incompact3d
ASKAP:
  tConvolve OpenMP - Degridding
  tConvolve OpenMP - Gridding
oneDNN:
  Deconvolution Batch shapes_1d - u8s8f32 - CPU
  IP Shapes 1D - f32 - CPU
TNN
Mlpack Benchmark
TNN
ASKAP
oneDNN:
  Convolution Batch Shapes Auto - f32 - CPU
  Convolution Batch Shapes Auto - u8s8f32 - CPU
  IP Shapes 3D - f32 - CPU
  IP Shapes 3D - u8s8f32 - CPU
WebP2 Image Encode
TNN
ECP-CANDLE
oneDNN:
  Deconvolution Batch shapes_3d - f32 - CPU
  Deconvolution Batch shapes_3d - u8s8f32 - CPU
WebP2 Image Encode