AMD Ryzen 7 7800X3D Linux Tests for a future article. AMD Ryzen 7 7800X3D 8-Core testing with a ASUS ROG CROSSHAIR X670E HERO (9927 BIOS) and AMD Radeon RX 7900 XTX on Ubuntu 23.04 via the Phoronix Test Suite.
HTML result view exported from: https://openbenchmarking.org/result/2304054-PTS-AMDRYZEN15&sro&grs .
AMD Ryzen 7 7800X3D Linux Processor Motherboard Chipset Memory Disk Graphics Audio Monitor Network OS Kernel Desktop Display Server OpenGL OpenCL Compiler File-System Screen Resolution a b AMD Ryzen 7 7800X3D 8-Core @ 4.20GHz (8 Cores / 16 Threads) ASUS ROG CROSSHAIR X670E HERO (9927 BIOS) AMD Device 14d8 32GB Western Digital WD_BLACK SN850X 1000GB AMD Radeon RX 7900 XTX (2304/1249MHz) AMD Device ab30 ASUS MG28U Intel I225-V + Intel Wi-Fi 6 AX210/AX211/AX411 Ubuntu 23.04 6.2.8-060208-generic (x86_64) GNOME Shell 44.0 X Server 1.21.1.7 + Wayland 4.6 Mesa 23.1.0-devel (git-de8b14f 2023-03-24 lunar-oibaf-ppa) (LLVM 15.0.7 DRM 3.49) OpenCL 2.1 AMD-APP (3513.0) GCC 12.2.0 ext4 3840x2160 OpenBenchmarking.org Kernel Details - Transparent Huge Pages: madvise Compiler Details - --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-12-Pa930Z/gcc-12-12.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-12-Pa930Z/gcc-12-12.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details - Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa601203 Java Details - OpenJDK Runtime Environment (build 17.0.6+10-Ubuntu-1ubuntu2) Python Details - Python 3.11.2 Security Details - itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected
AMD Ryzen 7 7800X3D Linux openfoam: drivaerFastback, Small Mesh Size - Mesh Time gnuradio: Five Back to Back FIR Filters askap: tConvolve OpenMP - Gridding askap: tConvolve MT - Degridding mt-dgemm: Sustained Floating-Point Rate askap: tConvolve MPI - Gridding onednn: Deconvolution Batch shapes_1d - f32 - CPU renaissance: ALS Movie Lens askap: tConvolve MPI - Degridding renaissance: Akka Unbalanced Cobwebbed Tree renaissance: Apache Spark Bayes askap: tConvolve MT - Gridding gnuradio: FM Deemphasis Filter apache: 200 renaissance: Finagle HTTP Requests clickhouse: 100M Rows Hits Dataset, Third Run clickhouse: 100M Rows Hits Dataset, Second Run apache: 500 apache: 100 renaissance: Genetic Algorithm Using Jenetics + Futures gnuradio: IIR Filter renaissance: Apache Spark PageRank numenta-nab: Relative Entropy numenta-nab: Windowed Gaussian numenta-nab: Earthgecko Skyline renaissance: Savina Reactors.IO numenta-nab: Bayesian Changepoint compress-7zip: Compression Rating renaissance: Scala Dotty memcached: 1:5 onednn: Deconvolution Batch shapes_1d - bf16bf16bf16 - CPU numenta-nab: Contextual Anomaly Detector OSE memcached: 1:10 cloverleaf: Lagrangian-Eulerian Hydrodynamics apache: 1000 onednn: Convolution Batch Shapes Auto - u8s8f32 - CPU renaissance: Apache Spark ALS gnuradio: FIR Filter openfoam: drivaerFastback, Medium Mesh Size - Mesh Time renaissance: Rand Forest onednn: IP Shapes 3D - u8s8f32 - CPU memcached: 1:100 build-godot: Time To Compile xmrig: Monero - 1M astcenc: Thorough compress-7zip: Decompression Rating gnuradio: Signal Source (Cosine) onednn: IP Shapes 1D - f32 - CPU onednn: Deconvolution Batch shapes_3d - u8s8f32 - CPU onednn: Convolution Batch Shapes Auto - f32 - CPU nginx: 1000 nginx: 500 tensorflow: CPU - 32 - ResNet-50 openfoam: drivaerFastback, Small Mesh Size - Execution Time build2: Time To Compile gnuradio: Hilbert Transform tensorflow: CPU - 64 - GoogLeNet onednn: IP Shapes 3D - bf16bf16bf16 - CPU onednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPU onednn: IP Shapes 1D - bf16bf16bf16 - CPU astcenc: Medium openfoam: drivaerFastback, Medium Mesh Size - Execution Time build-linux-kernel: defconfig tensorflow: CPU - 64 - ResNet-50 astcenc: Fast nginx: 100 nginx: 200 astcenc: Exhaustive onednn: Recurrent Neural Network Training - u8s8f32 - CPU tensorflow: CPU - 32 - AlexNet onednn: Deconvolution Batch shapes_3d - bf16bf16bf16 - CPU pennant: sedovbig amg: xmrig: Wownero - 1M pennant: leblancbig onednn: Recurrent Neural Network Inference - f32 - CPU onednn: Recurrent Neural Network Training - bf16bf16bf16 - CPU tensorflow: CPU - 32 - GoogLeNet blender: BMW27 - CPU-Only onednn: IP Shapes 1D - u8s8f32 - CPU numenta-nab: KNN CAD onednn: IP Shapes 3D - f32 - CPU onednn: Recurrent Neural Network Inference - u8s8f32 - CPU onednn: Deconvolution Batch shapes_3d - f32 - CPU tensorflow: CPU - 64 - AlexNet onednn: Convolution Batch Shapes Auto - bf16bf16bf16 - CPU clickhouse: 100M Rows Hits Dataset, First Run / Cold Cache onednn: Deconvolution Batch shapes_1d - u8s8f32 - CPU onednn: Recurrent Neural Network Training - f32 - CPU askap: Hogbom Clean OpenMP askap: tConvolve OpenMP - Degridding renaissance: In-Memory Database Shootout a b 27.199787 1821.2 8068.36 2252.83 6.112455 10495.8 5.51978 5903.6 8464.38 5680.7 900.0 1914.65 1113.8 193855.44 1879.6 278.90 274.60 185789.3 179311.78 953.8 500.9 1787.6 9.848 5.654 61.734 3153.6 17.331 112513 439.9 3272684.62 7.66355 25.456 2921694.2 39.39 182530.28 4.74584 1931.2 1391.1 222.26157 402.3 0.365467 2841542.11 300.001 9490.8 7.822 87240 5454.6 3.33782 1.01001 5.16302 93560.28 96954.32 33.57 169.69825 119.412 704.4 96.41 1.21056 1121.9 1.25071 63.3194 2121.7484 80.728 32.74 181.0022 97810.4 98163.47 0.8299 2227.52 177.73 2.47403 42.31317 406211800 10015.9 27.25888 1121 2227.01 101.24 104.47 0.620111 152.925 3.09967 1122.76 4.1026 209.51 2.04864 247.77 0.830924 2226.82 735.294 9861.33 29.932709 1715.3 7607.31 2159.2 6.373565 10933.2 5.32498 5716.0 8199.86 5860.4 873.8 1861.93 1141.3 189635.84 1843.8 273.75 270.69 183386.87 181582.13 942.4 506.6 1807.0 9.948 5.599 62.237 3128.9 17.196 113370 436.6 3250299.72 7.71613 25.286 2902278.7 39.15 183626.22 4.71913 1941.4 1398.4 223.41288 400.5 0.367087 2829008.81 298.68 9531.3 7.8537 86914 5472.7 3.34819 1.00696 5.14851 93822.98 96692.89 33.66 170.15311 119.155 702.9 96.61 1.20831 1123.68 1.25253 63.2296 2118.7619 80.833 32.78 181.1977 97914.05 98262.15 0.8305 2229.02 177.84 2.47255 42.33833 406014500 10011.1 27.24618 1121.34 2227.67 101.27 104.5 0.619937 152.884 3.09914 1122.58 4.10323 209.54 2.04844 247.75 0.830898 2226.86 735.294 9861.33 OpenBenchmarking.org
OpenFOAM Input: drivaerFastback, Small Mesh Size - Mesh Time OpenBenchmarking.org Seconds, Fewer Is Better OpenFOAM 10 Input: drivaerFastback, Small Mesh Size - Mesh Time a b 7 14 21 28 35 27.20 29.93 1. (CXX) g++ options: -std=c++14 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm
GNU Radio Test: Five Back to Back FIR Filters OpenBenchmarking.org MiB/s, More Is Better GNU Radio Test: Five Back to Back FIR Filters a b 400 800 1200 1600 2000 1821.2 1715.3 1. 3.10.5.1
ASKAP Test: tConvolve OpenMP - Gridding OpenBenchmarking.org Million Grid Points Per Second, More Is Better ASKAP 1.0 Test: tConvolve OpenMP - Gridding a b 2K 4K 6K 8K 10K 8068.36 7607.31 1. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp
ASKAP Test: tConvolve MT - Degridding OpenBenchmarking.org Million Grid Points Per Second, More Is Better ASKAP 1.0 Test: tConvolve MT - Degridding a b 500 1000 1500 2000 2500 2252.83 2159.20 1. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp
ACES DGEMM Sustained Floating-Point Rate OpenBenchmarking.org GFLOP/s, More Is Better ACES DGEMM 1.0 Sustained Floating-Point Rate a b 2 4 6 8 10 6.112455 6.373565 1. (CC) gcc options: -O3 -march=native -fopenmp
ASKAP Test: tConvolve MPI - Gridding OpenBenchmarking.org Mpix/sec, More Is Better ASKAP 1.0 Test: tConvolve MPI - Gridding a b 2K 4K 6K 8K 10K 10495.8 10933.2 1. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp
oneDNN Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU a b 1.242 2.484 3.726 4.968 6.21 5.51978 5.32498 MIN: 4.31 MIN: 4.32 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
Renaissance Test: ALS Movie Lens OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: ALS Movie Lens a b 1300 2600 3900 5200 6500 5903.6 5716.0 MIN: 5903.57 / MAX: 6389.29 MIN: 5715.97 / MAX: 6221.28
ASKAP Test: tConvolve MPI - Degridding OpenBenchmarking.org Mpix/sec, More Is Better ASKAP 1.0 Test: tConvolve MPI - Degridding a b 2K 4K 6K 8K 10K 8464.38 8199.86 1. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp
Renaissance Test: Akka Unbalanced Cobwebbed Tree OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Akka Unbalanced Cobwebbed Tree a b 1300 2600 3900 5200 6500 5680.7 5860.4 MIN: 4137.76 / MAX: 5680.74 MIN: 4368.85 / MAX: 5860.42
Renaissance Test: Apache Spark Bayes OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Apache Spark Bayes a b 200 400 600 800 1000 900.0 873.8 MIN: 657.2 / MAX: 900.04 MIN: 643.52
ASKAP Test: tConvolve MT - Gridding OpenBenchmarking.org Million Grid Points Per Second, More Is Better ASKAP 1.0 Test: tConvolve MT - Gridding a b 400 800 1200 1600 2000 1914.65 1861.93 1. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp
GNU Radio Test: FM Deemphasis Filter OpenBenchmarking.org MiB/s, More Is Better GNU Radio Test: FM Deemphasis Filter a b 200 400 600 800 1000 1113.8 1141.3 1. 3.10.5.1
Apache HTTP Server Concurrent Requests: 200 OpenBenchmarking.org Requests Per Second, More Is Better Apache HTTP Server 2.4.56 Concurrent Requests: 200 a b 40K 80K 120K 160K 200K 193855.44 189635.84 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
Renaissance Test: Finagle HTTP Requests OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Finagle HTTP Requests a b 400 800 1200 1600 2000 1879.6 1843.8 MIN: 1691.04 / MAX: 2039.24 MIN: 1681.09 / MAX: 2064.14
ClickHouse 100M Rows Hits Dataset, Third Run OpenBenchmarking.org Queries Per Minute, Geo Mean, More Is Better ClickHouse 22.12.3.5 100M Rows Hits Dataset, Third Run a b 60 120 180 240 300 278.90 273.75 MIN: 10.07 / MAX: 10000 MIN: 10.11 / MAX: 10000
ClickHouse 100M Rows Hits Dataset, Second Run OpenBenchmarking.org Queries Per Minute, Geo Mean, More Is Better ClickHouse 22.12.3.5 100M Rows Hits Dataset, Second Run a b 60 120 180 240 300 274.60 270.69 MIN: 10.07 / MAX: 10000 MIN: 10.07 / MAX: 8571.43
Apache HTTP Server Concurrent Requests: 500 OpenBenchmarking.org Requests Per Second, More Is Better Apache HTTP Server 2.4.56 Concurrent Requests: 500 a b 40K 80K 120K 160K 200K 185789.30 183386.87 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
Apache HTTP Server Concurrent Requests: 100 OpenBenchmarking.org Requests Per Second, More Is Better Apache HTTP Server 2.4.56 Concurrent Requests: 100 a b 40K 80K 120K 160K 200K 179311.78 181582.13 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
Renaissance Test: Genetic Algorithm Using Jenetics + Futures OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Genetic Algorithm Using Jenetics + Futures a b 200 400 600 800 1000 953.8 942.4 MIN: 940.04 / MAX: 966.17 MIN: 923.34 / MAX: 952.77
GNU Radio Test: IIR Filter OpenBenchmarking.org MiB/s, More Is Better GNU Radio Test: IIR Filter a b 110 220 330 440 550 500.9 506.6 1. 3.10.5.1
Renaissance Test: Apache Spark PageRank OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Apache Spark PageRank a b 400 800 1200 1600 2000 1787.6 1807.0 MIN: 1617.18 / MAX: 1862.42 MIN: 1679.68 / MAX: 1858.47
Numenta Anomaly Benchmark Detector: Relative Entropy OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: Relative Entropy a b 3 6 9 12 15 9.848 9.948
Numenta Anomaly Benchmark Detector: Windowed Gaussian OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: Windowed Gaussian a b 1.2722 2.5444 3.8166 5.0888 6.361 5.654 5.599
Numenta Anomaly Benchmark Detector: Earthgecko Skyline OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: Earthgecko Skyline a b 14 28 42 56 70 61.73 62.24
Renaissance Test: Savina Reactors.IO OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Savina Reactors.IO a b 700 1400 2100 2800 3500 3153.6 3128.9 MAX: 4292.86 MAX: 4313.77
Numenta Anomaly Benchmark Detector: Bayesian Changepoint OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: Bayesian Changepoint a b 4 8 12 16 20 17.33 17.20
7-Zip Compression Test: Compression Rating OpenBenchmarking.org MIPS, More Is Better 7-Zip Compression 22.01 Test: Compression Rating a b 20K 40K 60K 80K 100K 112513 113370 1. (CXX) g++ options: -lpthread -ldl -O2 -fPIC
Renaissance Test: Scala Dotty OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Scala Dotty a b 100 200 300 400 500 439.9 436.6 MIN: 366.56 / MAX: 788.38 MIN: 367.97 / MAX: 787.2
Memcached Set To Get Ratio: 1:5 OpenBenchmarking.org Ops/sec, More Is Better Memcached 1.6.19 Set To Get Ratio: 1:5 a b 700K 1400K 2100K 2800K 3500K 3272684.62 3250299.72 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
oneDNN Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU a b 2 4 6 8 10 7.66355 7.71613 MIN: 7.39 MIN: 7.42 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
Numenta Anomaly Benchmark Detector: Contextual Anomaly Detector OSE OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: Contextual Anomaly Detector OSE a b 6 12 18 24 30 25.46 25.29
Memcached Set To Get Ratio: 1:10 OpenBenchmarking.org Ops/sec, More Is Better Memcached 1.6.19 Set To Get Ratio: 1:10 a b 600K 1200K 1800K 2400K 3000K 2921694.2 2902278.7 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
CloverLeaf Lagrangian-Eulerian Hydrodynamics OpenBenchmarking.org Seconds, Fewer Is Better CloverLeaf Lagrangian-Eulerian Hydrodynamics a b 9 18 27 36 45 39.39 39.15 1. (F9X) gfortran options: -O3 -march=native -funroll-loops -fopenmp
Apache HTTP Server Concurrent Requests: 1000 OpenBenchmarking.org Requests Per Second, More Is Better Apache HTTP Server 2.4.56 Concurrent Requests: 1000 a b 40K 80K 120K 160K 200K 182530.28 183626.22 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
oneDNN Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU a b 1.0678 2.1356 3.2034 4.2712 5.339 4.74584 4.71913 MIN: 4.62 MIN: 4.64 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
Renaissance Test: Apache Spark ALS OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Apache Spark ALS a b 400 800 1200 1600 2000 1931.2 1941.4 MIN: 1827.2 / MAX: 2003.61 MIN: 1885.03 / MAX: 2000.37
GNU Radio Test: FIR Filter OpenBenchmarking.org MiB/s, More Is Better GNU Radio Test: FIR Filter a b 300 600 900 1200 1500 1391.1 1398.4 1. 3.10.5.1
OpenFOAM Input: drivaerFastback, Medium Mesh Size - Mesh Time OpenBenchmarking.org Seconds, Fewer Is Better OpenFOAM 10 Input: drivaerFastback, Medium Mesh Size - Mesh Time a b 50 100 150 200 250 222.26 223.41 1. (CXX) g++ options: -std=c++14 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm
Renaissance Test: Random Forest OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Random Forest a b 90 180 270 360 450 402.3 400.5 MIN: 373.43 / MAX: 460.71 MIN: 371.65 / MAX: 445.18
oneDNN Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU a b 0.0826 0.1652 0.2478 0.3304 0.413 0.365467 0.367087 MIN: 0.35 MIN: 0.35 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
Memcached Set To Get Ratio: 1:100 OpenBenchmarking.org Ops/sec, More Is Better Memcached 1.6.19 Set To Get Ratio: 1:100 a b 600K 1200K 1800K 2400K 3000K 2841542.11 2829008.81 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
Timed Godot Game Engine Compilation Time To Compile OpenBenchmarking.org Seconds, Fewer Is Better Timed Godot Game Engine Compilation 4.0 Time To Compile a b 70 140 210 280 350 300.00 298.68
Xmrig Variant: Monero - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.18.1 Variant: Monero - Hash Count: 1M a b 2K 4K 6K 8K 10K 9490.8 9531.3 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
ASTC Encoder Preset: Thorough OpenBenchmarking.org MT/s, More Is Better ASTC Encoder 4.0 Preset: Thorough a b 2 4 6 8 10 7.8220 7.8537 1. (CXX) g++ options: -O3 -flto -pthread
7-Zip Compression Test: Decompression Rating OpenBenchmarking.org MIPS, More Is Better 7-Zip Compression 22.01 Test: Decompression Rating a b 20K 40K 60K 80K 100K 87240 86914 1. (CXX) g++ options: -lpthread -ldl -O2 -fPIC
GNU Radio Test: Signal Source (Cosine) OpenBenchmarking.org MiB/s, More Is Better GNU Radio Test: Signal Source (Cosine) a b 1200 2400 3600 4800 6000 5454.6 5472.7 1. 3.10.5.1
oneDNN Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU a b 0.7533 1.5066 2.2599 3.0132 3.7665 3.33782 3.34819 MIN: 3.16 MIN: 3.14 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU a b 0.2273 0.4546 0.6819 0.9092 1.1365 1.01001 1.00696 MIN: 0.99 MIN: 0.98 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU a b 1.1617 2.3234 3.4851 4.6468 5.8085 5.16302 5.14851 MIN: 5.06 MIN: 5.06 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
nginx Connections: 1000 OpenBenchmarking.org Requests Per Second, More Is Better nginx 1.23.2 Connections: 1000 a b 20K 40K 60K 80K 100K 93560.28 93822.98 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
nginx Connections: 500 OpenBenchmarking.org Requests Per Second, More Is Better nginx 1.23.2 Connections: 500 a b 20K 40K 60K 80K 100K 96954.32 96692.89 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
TensorFlow Device: CPU - Batch Size: 32 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 32 - Model: ResNet-50 a b 8 16 24 32 40 33.57 33.66
OpenFOAM Input: drivaerFastback, Small Mesh Size - Execution Time OpenBenchmarking.org Seconds, Fewer Is Better OpenFOAM 10 Input: drivaerFastback, Small Mesh Size - Execution Time a b 40 80 120 160 200 169.70 170.15 1. (CXX) g++ options: -std=c++14 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm
Build2 Time To Compile OpenBenchmarking.org Seconds, Fewer Is Better Build2 0.15 Time To Compile a b 30 60 90 120 150 119.41 119.16
GNU Radio Test: Hilbert Transform OpenBenchmarking.org MiB/s, More Is Better GNU Radio Test: Hilbert Transform a b 150 300 450 600 750 704.4 702.9 1. 3.10.5.1
TensorFlow Device: CPU - Batch Size: 64 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 64 - Model: GoogLeNet a b 20 40 60 80 100 96.41 96.61
oneDNN Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU a b 0.2724 0.5448 0.8172 1.0896 1.362 1.21056 1.20831 MIN: 1.17 MIN: 1.17 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU a b 200 400 600 800 1000 1121.90 1123.68 MIN: 1117.18 MIN: 1118.61 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU a b 0.2818 0.5636 0.8454 1.1272 1.409 1.25071 1.25253 MIN: 1.23 MIN: 1.23 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
ASTC Encoder Preset: Medium OpenBenchmarking.org MT/s, More Is Better ASTC Encoder 4.0 Preset: Medium a b 14 28 42 56 70 63.32 63.23 1. (CXX) g++ options: -O3 -flto -pthread
OpenFOAM Input: drivaerFastback, Medium Mesh Size - Execution Time OpenBenchmarking.org Seconds, Fewer Is Better OpenFOAM 10 Input: drivaerFastback, Medium Mesh Size - Execution Time a b 500 1000 1500 2000 2500 2121.75 2118.76 1. (CXX) g++ options: -std=c++14 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm
Timed Linux Kernel Compilation Build: defconfig OpenBenchmarking.org Seconds, Fewer Is Better Timed Linux Kernel Compilation 6.1 Build: defconfig a b 20 40 60 80 100 80.73 80.83
TensorFlow Device: CPU - Batch Size: 64 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 64 - Model: ResNet-50 a b 8 16 24 32 40 32.74 32.78
ASTC Encoder Preset: Fast OpenBenchmarking.org MT/s, More Is Better ASTC Encoder 4.0 Preset: Fast a b 40 80 120 160 200 181.00 181.20 1. (CXX) g++ options: -O3 -flto -pthread
nginx Connections: 100 OpenBenchmarking.org Requests Per Second, More Is Better nginx 1.23.2 Connections: 100 a b 20K 40K 60K 80K 100K 97810.40 97914.05 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
nginx Connections: 200 OpenBenchmarking.org Requests Per Second, More Is Better nginx 1.23.2 Connections: 200 a b 20K 40K 60K 80K 100K 98163.47 98262.15 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
ASTC Encoder Preset: Exhaustive OpenBenchmarking.org MT/s, More Is Better ASTC Encoder 4.0 Preset: Exhaustive a b 0.1869 0.3738 0.5607 0.7476 0.9345 0.8299 0.8305 1. (CXX) g++ options: -O3 -flto -pthread
oneDNN Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU a b 500 1000 1500 2000 2500 2227.52 2229.02 MIN: 2223.04 MIN: 2222.68 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
TensorFlow Device: CPU - Batch Size: 32 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 32 - Model: AlexNet a b 40 80 120 160 200 177.73 177.84
oneDNN Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU a b 0.5567 1.1134 1.6701 2.2268 2.7835 2.47403 2.47255 MIN: 2.42 MIN: 2.41 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
Pennant Test: sedovbig OpenBenchmarking.org Hydro Cycle Time - Seconds, Fewer Is Better Pennant 1.0.1 Test: sedovbig a b 10 20 30 40 50 42.31 42.34 1. (CXX) g++ options: -fopenmp -lmpi_cxx -lmpi
Algebraic Multi-Grid Benchmark OpenBenchmarking.org Figure Of Merit, More Is Better Algebraic Multi-Grid Benchmark 1.2 a b 90M 180M 270M 360M 450M 406211800 406014500 1. (CC) gcc options: -lparcsr_ls -lparcsr_mv -lseq_mv -lIJ_mv -lkrylov -lHYPRE_utilities -lm -fopenmp -lmpi
Xmrig Variant: Wownero - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.18.1 Variant: Wownero - Hash Count: 1M a b 2K 4K 6K 8K 10K 10015.9 10011.1 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
Pennant Test: leblancbig OpenBenchmarking.org Hydro Cycle Time - Seconds, Fewer Is Better Pennant 1.0.1 Test: leblancbig a b 6 12 18 24 30 27.26 27.25 1. (CXX) g++ options: -fopenmp -lmpi_cxx -lmpi
oneDNN Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU a b 200 400 600 800 1000 1121.00 1121.34 MIN: 1117.71 MIN: 1117.68 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU a b 500 1000 1500 2000 2500 2227.01 2227.67 MIN: 2223.59 MIN: 2223.36 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
TensorFlow Device: CPU - Batch Size: 32 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 32 - Model: GoogLeNet a b 20 40 60 80 100 101.24 101.27
Blender Blend File: BMW27 - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 3.5 Blend File: BMW27 - Compute: CPU-Only a b 20 40 60 80 100 104.47 104.50
oneDNN Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU a b 0.1395 0.279 0.4185 0.558 0.6975 0.620111 0.619937 MIN: 0.61 MIN: 0.61 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
Numenta Anomaly Benchmark Detector: KNN CAD OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: KNN CAD a b 30 60 90 120 150 152.93 152.88
oneDNN Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU a b 0.6974 1.3948 2.0922 2.7896 3.487 3.09967 3.09914 MIN: 3.05 MIN: 3.05 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU a b 200 400 600 800 1000 1122.76 1122.58 MIN: 1118.01 MIN: 1118.65 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU a b 0.9232 1.8464 2.7696 3.6928 4.616 4.10260 4.10323 MIN: 4.05 MIN: 4.05 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
TensorFlow Device: CPU - Batch Size: 64 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 64 - Model: AlexNet a b 50 100 150 200 250 209.51 209.54
oneDNN Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU a b 0.4609 0.9218 1.3827 1.8436 2.3045 2.04864 2.04844 MIN: 2.01 MIN: 2.01 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
ClickHouse 100M Rows Hits Dataset, First Run / Cold Cache OpenBenchmarking.org Queries Per Minute, Geo Mean, More Is Better ClickHouse 22.12.3.5 100M Rows Hits Dataset, First Run / Cold Cache a b 50 100 150 200 250 247.77 247.75 MIN: 9.7 / MAX: 7500 MIN: 9.76 / MAX: 7500
oneDNN Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU a b 0.187 0.374 0.561 0.748 0.935 0.830924 0.830898 MIN: 0.82 MIN: 0.82 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU a b 500 1000 1500 2000 2500 2226.82 2226.86 MIN: 2223.91 MIN: 2222.67 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
ASKAP Test: Hogbom Clean OpenMP OpenBenchmarking.org Iterations Per Second, More Is Better ASKAP 1.0 Test: Hogbom Clean OpenMP a b 160 320 480 640 800 735.29 735.29 1. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp
ASKAP Test: tConvolve OpenMP - Degridding OpenBenchmarking.org Million Grid Points Per Second, More Is Better ASKAP 1.0 Test: tConvolve OpenMP - Degridding a b 2K 4K 6K 8K 10K 9861.33 9861.33 1. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp
Phoronix Test Suite v10.8.5