AMD Ryzen 7 7800X3D Linux AMD Ryzen 7 7800X3D 8-Core testing with a ASUS ROG CROSSHAIR X670E HERO (9927 BIOS) and AMD Radeon RX 7900 XTX on Ubuntu 23.04 via the Phoronix Test Suite.
HTML result view exported from: https://openbenchmarking.org/result/2304059-PTS-AMDRYZEN72&sor&grr .
AMD Ryzen 7 7800X3D Linux Processor Motherboard Chipset Memory Disk Graphics Audio Monitor Network OS Kernel Desktop Display Server OpenGL OpenCL Compiler File-System Screen Resolution a b AMD Ryzen 7 7800X3D 8-Core @ 4.20GHz (8 Cores / 16 Threads) ASUS ROG CROSSHAIR X670E HERO (9927 BIOS) AMD Device 14d8 32GB Western Digital WD_BLACK SN850X 1000GB AMD Radeon RX 7900 XTX (2304/1249MHz) AMD Device ab30 ASUS MG28U Intel I225-V + Intel Wi-Fi 6 AX210/AX211/AX411 Ubuntu 23.04 6.2.8-060208-generic (x86_64) GNOME Shell 44.0 X Server 1.21.1.7 + Wayland 4.6 Mesa 23.1.0-devel (git-de8b14f 2023-03-24 lunar-oibaf-ppa) (LLVM 15.0.7 DRM 3.49) OpenCL 2.1 AMD-APP (3513.0) GCC 12.2.0 ext4 3840x2160 OpenBenchmarking.org Kernel Details - Transparent Huge Pages: madvise Compiler Details - --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-12-Pa930Z/gcc-12-12.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-12-Pa930Z/gcc-12-12.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details - Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa601203 Java Details - OpenJDK Runtime Environment (build 17.0.6+10-Ubuntu-1ubuntu2) Python Details - Python 3.11.2 Security Details - itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected
AMD Ryzen 7 7800X3D Linux openfoam: drivaerFastback, Medium Mesh Size - Execution Time openfoam: drivaerFastback, Medium Mesh Size - Mesh Time gnuradio: Hilbert Transform gnuradio: FM Deemphasis Filter gnuradio: IIR Filter gnuradio: FIR Filter gnuradio: Signal Source (Cosine) gnuradio: Five Back to Back FIR Filters build-godot: Time To Compile tensorflow: CPU - 64 - ResNet-50 clickhouse: 100M Rows Hits Dataset, Third Run clickhouse: 100M Rows Hits Dataset, Second Run clickhouse: 100M Rows Hits Dataset, First Run / Cold Cache openfoam: drivaerFastback, Small Mesh Size - Execution Time openfoam: drivaerFastback, Small Mesh Size - Mesh Time numenta-nab: KNN CAD renaissance: ALS Movie Lens build2: Time To Compile renaissance: Akka Unbalanced Cobwebbed Tree tensorflow: CPU - 32 - ResNet-50 xmrig: Monero - 1M blender: BMW27 - CPU-Only xmrig: Wownero - 1M nginx: 500 nginx: 1000 apache: 500 apache: 1000 nginx: 200 nginx: 100 apache: 200 apache: 100 onednn: Recurrent Neural Network Training - f32 - CPU onednn: Recurrent Neural Network Training - u8s8f32 - CPU onednn: Recurrent Neural Network Training - bf16bf16bf16 - CPU build-linux-kernel: defconfig onednn: Recurrent Neural Network Inference - u8s8f32 - CPU onednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPU onednn: Recurrent Neural Network Inference - f32 - CPU tensorflow: CPU - 64 - GoogLeNet memcached: 1:5 memcached: 1:10 memcached: 1:100 renaissance: Apache Spark ALS numenta-nab: Earthgecko Skyline astcenc: Exhaustive renaissance: Apache Spark PageRank askap: tConvolve MT - Degridding askap: tConvolve MT - Gridding renaissance: Genetic Algorithm Using Jenetics + Futures pennant: sedovbig renaissance: Savina Reactors.IO cloverleaf: Lagrangian-Eulerian Hydrodynamics mt-dgemm: Sustained Floating-Point Rate tensorflow: CPU - 32 - GoogLeNet tensorflow: CPU - 64 - AlexNet renaissance: Scala Dotty astcenc: Thorough pennant: leblancbig renaissance: Apache Spark Bayes numenta-nab: Contextual Anomaly Detector OSE renaissance: Finagle HTTP Requests tensorflow: CPU - 32 - AlexNet renaissance: Rand Forest onednn: Deconvolution Batch shapes_1d - bf16bf16bf16 - CPU onednn: Deconvolution Batch shapes_1d - f32 - CPU onednn: Deconvolution Batch shapes_1d - u8s8f32 - CPU astcenc: Fast compress-7zip: Decompression Rating compress-7zip: Compression Rating numenta-nab: Bayesian Changepoint onednn: IP Shapes 1D - f32 - CPU onednn: IP Shapes 1D - bf16bf16bf16 - CPU onednn: IP Shapes 1D - u8s8f32 - CPU askap: tConvolve MPI - Gridding askap: tConvolve MPI - Degridding amg: askap: Hogbom Clean OpenMP astcenc: Medium numenta-nab: Relative Entropy onednn: IP Shapes 3D - bf16bf16bf16 - CPU onednn: IP Shapes 3D - f32 - CPU onednn: IP Shapes 3D - u8s8f32 - CPU onednn: Convolution Batch Shapes Auto - u8s8f32 - CPU onednn: Convolution Batch Shapes Auto - f32 - CPU onednn: Convolution Batch Shapes Auto - bf16bf16bf16 - CPU numenta-nab: Windowed Gaussian askap: tConvolve OpenMP - Degridding askap: tConvolve OpenMP - Gridding onednn: Deconvolution Batch shapes_3d - bf16bf16bf16 - CPU onednn: Deconvolution Batch shapes_3d - f32 - CPU onednn: Deconvolution Batch shapes_3d - u8s8f32 - CPU askap: tConvolve OpenCL a b 2121.7484 222.26157 704.4 1113.8 500.9 1391.1 5454.6 1821.2 300.001 32.74 278.90 274.60 247.77 169.69825 27.199787 152.925 5903.6 119.412 5680.7 33.57 9490.8 104.47 10015.9 96954.32 93560.28 185789.3 182530.28 98163.47 97810.4 193855.44 179311.78 2226.82 2227.52 2227.01 80.728 1122.76 1121.9 1121 96.41 3272684.62 2921694.2 2841542.11 1931.2 61.734 0.8299 1787.6 2252.83 1914.65 953.8 42.31317 3153.6 39.39 6.112455 101.24 209.51 439.9 7.822 27.25888 900.0 25.456 1879.6 177.73 402.3 7.66355 5.51978 0.830924 181.0022 87240 112513 17.331 3.33782 1.25071 0.620111 10495.8 8464.38 406211800 735.294 63.3194 9.848 1.21056 3.09967 0.365467 4.74584 5.16302 2.04864 5.654 9861.33 8068.36 2.47403 4.1026 1.01001 2118.7619 223.41288 702.9 1141.3 506.6 1398.4 5472.7 1715.3 298.68 32.78 273.75 270.69 247.75 170.15311 29.932709 152.884 5716.0 119.155 5860.4 33.66 9531.3 104.5 10011.1 96692.89 93822.98 183386.87 183626.22 98262.15 97914.05 189635.84 181582.13 2226.86 2229.02 2227.67 80.833 1122.58 1123.68 1121.34 96.61 3250299.72 2902278.7 2829008.81 1941.4 62.237 0.8305 1807.0 2159.2 1861.93 942.4 42.33833 3128.9 39.15 6.373565 101.27 209.54 436.6 7.8537 27.24618 873.8 25.286 1843.8 177.84 400.5 7.71613 5.32498 0.830898 181.1977 86914 113370 17.196 3.34819 1.25253 0.619937 10933.2 8199.86 406014500 735.294 63.2296 9.948 1.20831 3.09914 0.367087 4.71913 5.14851 2.04844 5.599 9861.33 7607.31 2.47255 4.10323 1.00696 OpenBenchmarking.org
OpenFOAM Input: drivaerFastback, Medium Mesh Size - Execution Time OpenBenchmarking.org Seconds, Fewer Is Better OpenFOAM 10 Input: drivaerFastback, Medium Mesh Size - Execution Time b a 500 1000 1500 2000 2500 2118.76 2121.75 1. (CXX) g++ options: -std=c++14 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm
OpenFOAM Input: drivaerFastback, Medium Mesh Size - Mesh Time OpenBenchmarking.org Seconds, Fewer Is Better OpenFOAM 10 Input: drivaerFastback, Medium Mesh Size - Mesh Time a b 50 100 150 200 250 222.26 223.41 1. (CXX) g++ options: -std=c++14 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm
GNU Radio Test: Hilbert Transform OpenBenchmarking.org MiB/s, More Is Better GNU Radio Test: Hilbert Transform a b 150 300 450 600 750 704.4 702.9 1. 3.10.5.1
GNU Radio Test: FM Deemphasis Filter OpenBenchmarking.org MiB/s, More Is Better GNU Radio Test: FM Deemphasis Filter b a 200 400 600 800 1000 1141.3 1113.8 1. 3.10.5.1
GNU Radio Test: IIR Filter OpenBenchmarking.org MiB/s, More Is Better GNU Radio Test: IIR Filter b a 110 220 330 440 550 506.6 500.9 1. 3.10.5.1
GNU Radio Test: FIR Filter OpenBenchmarking.org MiB/s, More Is Better GNU Radio Test: FIR Filter b a 300 600 900 1200 1500 1398.4 1391.1 1. 3.10.5.1
GNU Radio Test: Signal Source (Cosine) OpenBenchmarking.org MiB/s, More Is Better GNU Radio Test: Signal Source (Cosine) b a 1200 2400 3600 4800 6000 5472.7 5454.6 1. 3.10.5.1
GNU Radio Test: Five Back to Back FIR Filters OpenBenchmarking.org MiB/s, More Is Better GNU Radio Test: Five Back to Back FIR Filters a b 400 800 1200 1600 2000 1821.2 1715.3 1. 3.10.5.1
Timed Godot Game Engine Compilation Time To Compile OpenBenchmarking.org Seconds, Fewer Is Better Timed Godot Game Engine Compilation 4.0 Time To Compile b a 70 140 210 280 350 298.68 300.00
TensorFlow Device: CPU - Batch Size: 64 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 64 - Model: ResNet-50 b a 8 16 24 32 40 32.78 32.74
ClickHouse 100M Rows Hits Dataset, Third Run OpenBenchmarking.org Queries Per Minute, Geo Mean, More Is Better ClickHouse 22.12.3.5 100M Rows Hits Dataset, Third Run a b 60 120 180 240 300 278.90 273.75 MIN: 10.07 / MAX: 10000 MIN: 10.11 / MAX: 10000
ClickHouse 100M Rows Hits Dataset, Second Run OpenBenchmarking.org Queries Per Minute, Geo Mean, More Is Better ClickHouse 22.12.3.5 100M Rows Hits Dataset, Second Run a b 60 120 180 240 300 274.60 270.69 MIN: 10.07 / MAX: 10000 MIN: 10.07 / MAX: 8571.43
ClickHouse 100M Rows Hits Dataset, First Run / Cold Cache OpenBenchmarking.org Queries Per Minute, Geo Mean, More Is Better ClickHouse 22.12.3.5 100M Rows Hits Dataset, First Run / Cold Cache a b 50 100 150 200 250 247.77 247.75 MIN: 9.7 / MAX: 7500 MIN: 9.76 / MAX: 7500
OpenFOAM Input: drivaerFastback, Small Mesh Size - Execution Time OpenBenchmarking.org Seconds, Fewer Is Better OpenFOAM 10 Input: drivaerFastback, Small Mesh Size - Execution Time a b 40 80 120 160 200 169.70 170.15 1. (CXX) g++ options: -std=c++14 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm
OpenFOAM Input: drivaerFastback, Small Mesh Size - Mesh Time OpenBenchmarking.org Seconds, Fewer Is Better OpenFOAM 10 Input: drivaerFastback, Small Mesh Size - Mesh Time a b 7 14 21 28 35 27.20 29.93 1. (CXX) g++ options: -std=c++14 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm
Numenta Anomaly Benchmark Detector: KNN CAD OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: KNN CAD b a 30 60 90 120 150 152.88 152.93
Renaissance Test: ALS Movie Lens OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: ALS Movie Lens b a 1300 2600 3900 5200 6500 5716.0 5903.6 MIN: 5715.97 / MAX: 6221.28 MIN: 5903.57 / MAX: 6389.29
Build2 Time To Compile OpenBenchmarking.org Seconds, Fewer Is Better Build2 0.15 Time To Compile b a 30 60 90 120 150 119.16 119.41
Renaissance Test: Akka Unbalanced Cobwebbed Tree OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Akka Unbalanced Cobwebbed Tree a b 1300 2600 3900 5200 6500 5680.7 5860.4 MIN: 4137.76 / MAX: 5680.74 MIN: 4368.85 / MAX: 5860.42
TensorFlow Device: CPU - Batch Size: 32 - Model: ResNet-50 OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 32 - Model: ResNet-50 b a 8 16 24 32 40 33.66 33.57
Xmrig Variant: Monero - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.18.1 Variant: Monero - Hash Count: 1M b a 2K 4K 6K 8K 10K 9531.3 9490.8 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
Blender Blend File: BMW27 - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 3.5 Blend File: BMW27 - Compute: CPU-Only a b 20 40 60 80 100 104.47 104.50
Xmrig Variant: Wownero - Hash Count: 1M OpenBenchmarking.org H/s, More Is Better Xmrig 6.18.1 Variant: Wownero - Hash Count: 1M a b 2K 4K 6K 8K 10K 10015.9 10011.1 1. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc
nginx Connections: 500 OpenBenchmarking.org Requests Per Second, More Is Better nginx 1.23.2 Connections: 500 a b 20K 40K 60K 80K 100K 96954.32 96692.89 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
nginx Connections: 1000 OpenBenchmarking.org Requests Per Second, More Is Better nginx 1.23.2 Connections: 1000 b a 20K 40K 60K 80K 100K 93822.98 93560.28 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
Apache HTTP Server Concurrent Requests: 500 OpenBenchmarking.org Requests Per Second, More Is Better Apache HTTP Server 2.4.56 Concurrent Requests: 500 a b 40K 80K 120K 160K 200K 185789.30 183386.87 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
Apache HTTP Server Concurrent Requests: 1000 OpenBenchmarking.org Requests Per Second, More Is Better Apache HTTP Server 2.4.56 Concurrent Requests: 1000 b a 40K 80K 120K 160K 200K 183626.22 182530.28 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
nginx Connections: 200 OpenBenchmarking.org Requests Per Second, More Is Better nginx 1.23.2 Connections: 200 b a 20K 40K 60K 80K 100K 98262.15 98163.47 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
nginx Connections: 100 OpenBenchmarking.org Requests Per Second, More Is Better nginx 1.23.2 Connections: 100 b a 20K 40K 60K 80K 100K 97914.05 97810.40 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
Apache HTTP Server Concurrent Requests: 200 OpenBenchmarking.org Requests Per Second, More Is Better Apache HTTP Server 2.4.56 Concurrent Requests: 200 a b 40K 80K 120K 160K 200K 193855.44 189635.84 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
Apache HTTP Server Concurrent Requests: 100 OpenBenchmarking.org Requests Per Second, More Is Better Apache HTTP Server 2.4.56 Concurrent Requests: 100 b a 40K 80K 120K 160K 200K 181582.13 179311.78 1. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2
oneDNN Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU a b 500 1000 1500 2000 2500 2226.82 2226.86 MIN: 2223.91 MIN: 2222.67 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU a b 500 1000 1500 2000 2500 2227.52 2229.02 MIN: 2223.04 MIN: 2222.68 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU a b 500 1000 1500 2000 2500 2227.01 2227.67 MIN: 2223.59 MIN: 2223.36 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
Timed Linux Kernel Compilation Build: defconfig OpenBenchmarking.org Seconds, Fewer Is Better Timed Linux Kernel Compilation 6.1 Build: defconfig a b 20 40 60 80 100 80.73 80.83
oneDNN Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU b a 200 400 600 800 1000 1122.58 1122.76 MIN: 1118.65 MIN: 1118.01 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU a b 200 400 600 800 1000 1121.90 1123.68 MIN: 1117.18 MIN: 1118.61 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU a b 200 400 600 800 1000 1121.00 1121.34 MIN: 1117.71 MIN: 1117.68 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
TensorFlow Device: CPU - Batch Size: 64 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 64 - Model: GoogLeNet b a 20 40 60 80 100 96.61 96.41
Memcached Set To Get Ratio: 1:5 OpenBenchmarking.org Ops/sec, More Is Better Memcached 1.6.19 Set To Get Ratio: 1:5 a b 700K 1400K 2100K 2800K 3500K 3272684.62 3250299.72 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
Memcached Set To Get Ratio: 1:10 OpenBenchmarking.org Ops/sec, More Is Better Memcached 1.6.19 Set To Get Ratio: 1:10 a b 600K 1200K 1800K 2400K 3000K 2921694.2 2902278.7 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
Memcached Set To Get Ratio: 1:100 OpenBenchmarking.org Ops/sec, More Is Better Memcached 1.6.19 Set To Get Ratio: 1:100 a b 600K 1200K 1800K 2400K 3000K 2841542.11 2829008.81 1. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre
Renaissance Test: Apache Spark ALS OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Apache Spark ALS a b 400 800 1200 1600 2000 1931.2 1941.4 MIN: 1827.2 / MAX: 2003.61 MIN: 1885.03 / MAX: 2000.37
Numenta Anomaly Benchmark Detector: Earthgecko Skyline OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: Earthgecko Skyline a b 14 28 42 56 70 61.73 62.24
ASTC Encoder Preset: Exhaustive OpenBenchmarking.org MT/s, More Is Better ASTC Encoder 4.0 Preset: Exhaustive b a 0.1869 0.3738 0.5607 0.7476 0.9345 0.8305 0.8299 1. (CXX) g++ options: -O3 -flto -pthread
Renaissance Test: Apache Spark PageRank OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Apache Spark PageRank a b 400 800 1200 1600 2000 1787.6 1807.0 MIN: 1617.18 / MAX: 1862.42 MIN: 1679.68 / MAX: 1858.47
ASKAP Test: tConvolve MT - Degridding OpenBenchmarking.org Million Grid Points Per Second, More Is Better ASKAP 1.0 Test: tConvolve MT - Degridding a b 500 1000 1500 2000 2500 2252.83 2159.20 1. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp
ASKAP Test: tConvolve MT - Gridding OpenBenchmarking.org Million Grid Points Per Second, More Is Better ASKAP 1.0 Test: tConvolve MT - Gridding a b 400 800 1200 1600 2000 1914.65 1861.93 1. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp
Renaissance Test: Genetic Algorithm Using Jenetics + Futures OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Genetic Algorithm Using Jenetics + Futures b a 200 400 600 800 1000 942.4 953.8 MIN: 923.34 / MAX: 952.77 MIN: 940.04 / MAX: 966.17
Pennant Test: sedovbig OpenBenchmarking.org Hydro Cycle Time - Seconds, Fewer Is Better Pennant 1.0.1 Test: sedovbig a b 10 20 30 40 50 42.31 42.34 1. (CXX) g++ options: -fopenmp -lmpi_cxx -lmpi
Renaissance Test: Savina Reactors.IO OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Savina Reactors.IO b a 700 1400 2100 2800 3500 3128.9 3153.6 MAX: 4313.77 MAX: 4292.86
CloverLeaf Lagrangian-Eulerian Hydrodynamics OpenBenchmarking.org Seconds, Fewer Is Better CloverLeaf Lagrangian-Eulerian Hydrodynamics b a 9 18 27 36 45 39.15 39.39 1. (F9X) gfortran options: -O3 -march=native -funroll-loops -fopenmp
ACES DGEMM Sustained Floating-Point Rate OpenBenchmarking.org GFLOP/s, More Is Better ACES DGEMM 1.0 Sustained Floating-Point Rate b a 2 4 6 8 10 6.373565 6.112455 1. (CC) gcc options: -O3 -march=native -fopenmp
TensorFlow Device: CPU - Batch Size: 32 - Model: GoogLeNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 32 - Model: GoogLeNet b a 20 40 60 80 100 101.27 101.24
TensorFlow Device: CPU - Batch Size: 64 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 64 - Model: AlexNet b a 50 100 150 200 250 209.54 209.51
Renaissance Test: Scala Dotty OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Scala Dotty b a 100 200 300 400 500 436.6 439.9 MIN: 367.97 / MAX: 787.2 MIN: 366.56 / MAX: 788.38
ASTC Encoder Preset: Thorough OpenBenchmarking.org MT/s, More Is Better ASTC Encoder 4.0 Preset: Thorough b a 2 4 6 8 10 7.8537 7.8220 1. (CXX) g++ options: -O3 -flto -pthread
Pennant Test: leblancbig OpenBenchmarking.org Hydro Cycle Time - Seconds, Fewer Is Better Pennant 1.0.1 Test: leblancbig b a 6 12 18 24 30 27.25 27.26 1. (CXX) g++ options: -fopenmp -lmpi_cxx -lmpi
Renaissance Test: Apache Spark Bayes OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Apache Spark Bayes b a 200 400 600 800 1000 873.8 900.0 MIN: 643.52 MIN: 657.2 / MAX: 900.04
Numenta Anomaly Benchmark Detector: Contextual Anomaly Detector OSE OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: Contextual Anomaly Detector OSE b a 6 12 18 24 30 25.29 25.46
Renaissance Test: Finagle HTTP Requests OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Finagle HTTP Requests b a 400 800 1200 1600 2000 1843.8 1879.6 MIN: 1681.09 / MAX: 2064.14 MIN: 1691.04 / MAX: 2039.24
TensorFlow Device: CPU - Batch Size: 32 - Model: AlexNet OpenBenchmarking.org images/sec, More Is Better TensorFlow 2.12 Device: CPU - Batch Size: 32 - Model: AlexNet b a 40 80 120 160 200 177.84 177.73
Renaissance Test: Random Forest OpenBenchmarking.org ms, Fewer Is Better Renaissance 0.14 Test: Random Forest b a 90 180 270 360 450 400.5 402.3 MIN: 371.65 / MAX: 445.18 MIN: 373.43 / MAX: 460.71
oneDNN Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU a b 2 4 6 8 10 7.66355 7.71613 MIN: 7.39 MIN: 7.42 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU b a 1.242 2.484 3.726 4.968 6.21 5.32498 5.51978 MIN: 4.32 MIN: 4.31 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU b a 0.187 0.374 0.561 0.748 0.935 0.830898 0.830924 MIN: 0.82 MIN: 0.82 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
ASTC Encoder Preset: Fast OpenBenchmarking.org MT/s, More Is Better ASTC Encoder 4.0 Preset: Fast b a 40 80 120 160 200 181.20 181.00 1. (CXX) g++ options: -O3 -flto -pthread
7-Zip Compression Test: Decompression Rating OpenBenchmarking.org MIPS, More Is Better 7-Zip Compression 22.01 Test: Decompression Rating a b 20K 40K 60K 80K 100K 87240 86914 1. (CXX) g++ options: -lpthread -ldl -O2 -fPIC
7-Zip Compression Test: Compression Rating OpenBenchmarking.org MIPS, More Is Better 7-Zip Compression 22.01 Test: Compression Rating b a 20K 40K 60K 80K 100K 113370 112513 1. (CXX) g++ options: -lpthread -ldl -O2 -fPIC
Numenta Anomaly Benchmark Detector: Bayesian Changepoint OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: Bayesian Changepoint b a 4 8 12 16 20 17.20 17.33
oneDNN Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU a b 0.7533 1.5066 2.2599 3.0132 3.7665 3.33782 3.34819 MIN: 3.16 MIN: 3.14 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU a b 0.2818 0.5636 0.8454 1.1272 1.409 1.25071 1.25253 MIN: 1.23 MIN: 1.23 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU b a 0.1395 0.279 0.4185 0.558 0.6975 0.619937 0.620111 MIN: 0.61 MIN: 0.61 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
ASKAP Test: tConvolve MPI - Gridding OpenBenchmarking.org Mpix/sec, More Is Better ASKAP 1.0 Test: tConvolve MPI - Gridding b a 2K 4K 6K 8K 10K 10933.2 10495.8 1. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp
ASKAP Test: tConvolve MPI - Degridding OpenBenchmarking.org Mpix/sec, More Is Better ASKAP 1.0 Test: tConvolve MPI - Degridding a b 2K 4K 6K 8K 10K 8464.38 8199.86 1. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp
Algebraic Multi-Grid Benchmark OpenBenchmarking.org Figure Of Merit, More Is Better Algebraic Multi-Grid Benchmark 1.2 a b 90M 180M 270M 360M 450M 406211800 406014500 1. (CC) gcc options: -lparcsr_ls -lparcsr_mv -lseq_mv -lIJ_mv -lkrylov -lHYPRE_utilities -lm -fopenmp -lmpi
ASKAP Test: Hogbom Clean OpenMP OpenBenchmarking.org Iterations Per Second, More Is Better ASKAP 1.0 Test: Hogbom Clean OpenMP b a 160 320 480 640 800 735.29 735.29 1. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp
ASTC Encoder Preset: Medium OpenBenchmarking.org MT/s, More Is Better ASTC Encoder 4.0 Preset: Medium a b 14 28 42 56 70 63.32 63.23 1. (CXX) g++ options: -O3 -flto -pthread
Numenta Anomaly Benchmark Detector: Relative Entropy OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: Relative Entropy a b 3 6 9 12 15 9.848 9.948
oneDNN Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU b a 0.2724 0.5448 0.8172 1.0896 1.362 1.20831 1.21056 MIN: 1.17 MIN: 1.17 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU b a 0.6974 1.3948 2.0922 2.7896 3.487 3.09914 3.09967 MIN: 3.05 MIN: 3.05 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU a b 0.0826 0.1652 0.2478 0.3304 0.413 0.365467 0.367087 MIN: 0.35 MIN: 0.35 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU b a 1.0678 2.1356 3.2034 4.2712 5.339 4.71913 4.74584 MIN: 4.64 MIN: 4.62 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU b a 1.1617 2.3234 3.4851 4.6468 5.8085 5.14851 5.16302 MIN: 5.06 MIN: 5.06 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU b a 0.4609 0.9218 1.3827 1.8436 2.3045 2.04844 2.04864 MIN: 2.01 MIN: 2.01 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
Numenta Anomaly Benchmark Detector: Windowed Gaussian OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: Windowed Gaussian b a 1.2722 2.5444 3.8166 5.0888 6.361 5.599 5.654
ASKAP Test: tConvolve OpenMP - Degridding OpenBenchmarking.org Million Grid Points Per Second, More Is Better ASKAP 1.0 Test: tConvolve OpenMP - Degridding b a 2K 4K 6K 8K 10K 9861.33 9861.33 1. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp
ASKAP Test: tConvolve OpenMP - Gridding OpenBenchmarking.org Million Grid Points Per Second, More Is Better ASKAP 1.0 Test: tConvolve OpenMP - Gridding a b 2K 4K 6K 8K 10K 8068.36 7607.31 1. (CXX) g++ options: -O3 -fstrict-aliasing -fopenmp
oneDNN Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU b a 0.5567 1.1134 1.6701 2.2268 2.7835 2.47255 2.47403 MIN: 2.41 MIN: 2.42 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU a b 0.9232 1.8464 2.7696 3.6928 4.616 4.10260 4.10323 MIN: 4.05 MIN: 4.05 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.1 Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU b a 0.2273 0.4546 0.6819 0.9092 1.1365 1.00696 1.01001 MIN: 0.98 MIN: 0.99 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
Phoronix Test Suite v10.8.5