AMD Ryzen 7 7840U testing with a PHX Ray_PEU (V1.04 BIOS) and AMD Phoenix1 512MB on Ubuntu 23.10 via the Phoronix Test Suite.
Compare your own system(s) to this result file with the
Phoronix Test Suite by running the command:
phoronix-test-suite benchmark 2310248-NE-DD557351896 dd - Phoronix Test Suite dd AMD Ryzen 7 7840U testing with a PHX Ray_PEU (V1.04 BIOS) and AMD Phoenix1 512MB on Ubuntu 23.10 via the Phoronix Test Suite.
HTML result view exported from: https://openbenchmarking.org/result/2310248-NE-DD557351896&sor&grr .
dd Processor Motherboard Chipset Memory Disk Graphics Audio Network OS Kernel Desktop Display Server OpenGL Compiler File-System Screen Resolution a b c AMD Ryzen 7 7840U @ 5.29GHz (8 Cores / 16 Threads) PHX Ray_PEU (V1.04 BIOS) AMD Device 14e8 16GB 1024GB Micron_3400_MTFDKBA1T0TFH AMD Phoenix1 512MB (2700/800MHz) AMD Rembrandt Radeon HD Audio MEDIATEK MT7922 802.11ax PCI Ubuntu 23.10 6.5.0-with-patch2 (x86_64) GNOME Shell 45.0 X Server + Wayland 4.6 Mesa 23.3~git2309080600.fd297e~oibaf~m (git-fd297ec 2023-09-08 mantic-oibaf-ppa) (LLVM 15.0.7 DRM 3.54) GCC 13.2.0 ext4 3200x2000 OpenBenchmarking.org Kernel Details - Transparent Huge Pages: madvise Compiler Details - a: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-13-4ZOwl4/gcc-13-13.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-13-4ZOwl4/gcc-13-13.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - b: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-13-4ZOwl4/gcc-13-13.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-13-4ZOwl4/gcc-13-13.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - c: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details - Scaling Governor: amd-pstate-epp powersave (EPP: balance_performance) - CPU Microcode: 0xa704101 Security Details - gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_rstack_overflow: Mitigation of safe RET no microcode + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected
dd openvkl: vklBenchmarkCPU ISPC openvkl: vklBenchmarkCPU Scalar easywave: e2Asean Grid + BengkuluSept2007 Source - 2400 oidn: RTLightmap.hdr.4096x4096 - CPU-Only easywave: e2Asean Grid + BengkuluSept2007 Source - 1200 oidn: RT.hdr_alb_nrm.3840x2160 - CPU-Only oidn: RT.ldr_alb_nrm.3840x2160 - CPU-Only embree: Pathtracer - Crown embree: Pathtracer - Asian Dragon Obj embree: Pathtracer ISPC - Crown embree: Pathtracer ISPC - Asian Dragon Obj onednn: Recurrent Neural Network Training - bf16bf16bf16 - CPU onednn: Recurrent Neural Network Training - u8s8f32 - CPU onednn: Recurrent Neural Network Training - f32 - CPU embree: Pathtracer - Asian Dragon onednn: Recurrent Neural Network Inference - u8s8f32 - CPU onednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPU onednn: Recurrent Neural Network Inference - f32 - CPU embree: Pathtracer ISPC - Asian Dragon quantlib: Multi-Threaded quantlib: Single-Threaded onednn: Deconvolution Batch shapes_1d - bf16bf16bf16 - CPU onednn: Deconvolution Batch shapes_1d - u8s8f32 - CPU onednn: Deconvolution Batch shapes_1d - f32 - CPU onednn: IP Shapes 1D - f32 - CPU onednn: IP Shapes 1D - bf16bf16bf16 - CPU onednn: IP Shapes 1D - u8s8f32 - CPU easywave: e2Asean Grid + BengkuluSept2007 Source - 240 onednn: IP Shapes 3D - bf16bf16bf16 - CPU onednn: IP Shapes 3D - f32 - CPU onednn: IP Shapes 3D - u8s8f32 - CPU onednn: Convolution Batch Shapes Auto - f32 - CPU onednn: Convolution Batch Shapes Auto - u8s8f32 - CPU onednn: Convolution Batch Shapes Auto - bf16bf16bf16 - CPU onednn: Deconvolution Batch shapes_3d - f32 - CPU onednn: Deconvolution Batch shapes_3d - bf16bf16bf16 - CPU onednn: Deconvolution Batch shapes_3d - u8s8f32 - CPU a b c 170 67 584.728 0.11 236.372 0.22 0.22 7.1781 7.8348 7.4556 8.1114 4956.7 4919.23 4898.17 8.6173 2506.07 2520.79 2514.18 9.5152 28605 4133 9.93332 1.52901 10.8682 8.86916 2.75717 1.2811 11.979 3.79236 5.6112 2.22678 15.0173 14.2507 6.76831 7.60937 5.17232 1.92455 171 67 585.354 0.11 235.936 0.22 0.22 7.2714 7.8132 7.4331 8.0696 4925.48 4965.28 4967.81 8.764 2555.28 2524.94 2555.43 9.5178 28768.7 4229.3 9.73785 1.67745 10.4152 8.99248 2.75383 1.25823 11.929 3.75671 5.57229 2.22532 14.9408 14.5776 6.74215 7.675 5.22381 2.01069 171 69 579.986 0.11 236.144 0.22 0.22 7.1882 7.7511 7.4151 8.0388 4574.04 4376.69 3876.53 8.733 2320.01 2382.55 2127.79 9.466 28299.9 4074.6 8.70755 1.29785 9.6233 8.66246 2.42519 1.1559 11.981 3.62443 5.59833 2.1978 15.0946 14.2659 6.72684 6.36614 4.42411 1.55587 OpenBenchmarking.org
OpenVKL Benchmark: vklBenchmarkCPU ISPC OpenBenchmarking.org Items / Sec, More Is Better OpenVKL 2.0.0 Benchmark: vklBenchmarkCPU ISPC c b a 40 80 120 160 200 171 171 170 MIN: 12 / MAX: 2390 MIN: 12 / MAX: 2399 MIN: 12 / MAX: 2389
OpenVKL Benchmark: vklBenchmarkCPU Scalar OpenBenchmarking.org Items / Sec, More Is Better OpenVKL 2.0.0 Benchmark: vklBenchmarkCPU Scalar c b a 15 30 45 60 75 69 67 67 MIN: 5 / MAX: 1266 MIN: 5 / MAX: 1194 MIN: 5 / MAX: 1197
easyWave Input: e2Asean Grid + BengkuluSept2007 Source - Time: 2400 OpenBenchmarking.org Seconds, Fewer Is Better easyWave r34 Input: e2Asean Grid + BengkuluSept2007 Source - Time: 2400 c a b 130 260 390 520 650 579.99 584.73 585.35 1. (CXX) g++ options: -O3 -fopenmp
Intel Open Image Denoise Run: RTLightmap.hdr.4096x4096 - Device: CPU-Only OpenBenchmarking.org Images / Sec, More Is Better Intel Open Image Denoise 2.1 Run: RTLightmap.hdr.4096x4096 - Device: CPU-Only c b a 0.0248 0.0496 0.0744 0.0992 0.124 0.11 0.11 0.11
easyWave Input: e2Asean Grid + BengkuluSept2007 Source - Time: 1200 OpenBenchmarking.org Seconds, Fewer Is Better easyWave r34 Input: e2Asean Grid + BengkuluSept2007 Source - Time: 1200 b c a 50 100 150 200 250 235.94 236.14 236.37 1. (CXX) g++ options: -O3 -fopenmp
Intel Open Image Denoise Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Only OpenBenchmarking.org Images / Sec, More Is Better Intel Open Image Denoise 2.1 Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Only c b a 0.0495 0.099 0.1485 0.198 0.2475 0.22 0.22 0.22
Intel Open Image Denoise Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Only OpenBenchmarking.org Images / Sec, More Is Better Intel Open Image Denoise 2.1 Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Only c b a 0.0495 0.099 0.1485 0.198 0.2475 0.22 0.22 0.22
Embree Binary: Pathtracer - Model: Crown OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer - Model: Crown b c a 2 4 6 8 10 7.2714 7.1882 7.1781 MIN: 7.2 / MAX: 7.51 MIN: 7.11 / MAX: 7.5 MIN: 7.09 / MAX: 7.6
Embree Binary: Pathtracer - Model: Asian Dragon Obj OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer - Model: Asian Dragon Obj a b c 2 4 6 8 10 7.8348 7.8132 7.7511 MIN: 7.8 / MAX: 7.98 MIN: 7.74 / MAX: 8.1 MIN: 7.7 / MAX: 7.89
Embree Binary: Pathtracer ISPC - Model: Crown OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer ISPC - Model: Crown a b c 2 4 6 8 10 7.4556 7.4331 7.4151 MIN: 7.37 / MAX: 7.79 MIN: 7.34 / MAX: 7.7 MIN: 7.34 / MAX: 7.69
Embree Binary: Pathtracer ISPC - Model: Asian Dragon Obj OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer ISPC - Model: Asian Dragon Obj a b c 2 4 6 8 10 8.1114 8.0696 8.0388 MIN: 8.07 / MAX: 8.27 MIN: 7.97 / MAX: 8.36 MIN: 7.99 / MAX: 8.16
oneDNN Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU c b a 1100 2200 3300 4400 5500 4574.04 4925.48 4956.70 MIN: 4522.29 MIN: 4884.66 MIN: 4902.84 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU c a b 1100 2200 3300 4400 5500 4376.69 4919.23 4965.28 MIN: 4320.61 MIN: 4873.66 MIN: 4916 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU c a b 1100 2200 3300 4400 5500 3876.53 4898.17 4967.81 MIN: 3780.33 MIN: 4855.43 MIN: 4924.41 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
Embree Binary: Pathtracer - Model: Asian Dragon OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer - Model: Asian Dragon b c a 2 4 6 8 10 8.7640 8.7330 8.6173 MIN: 8.67 / MAX: 9.04 MIN: 8.64 / MAX: 9.06 MIN: 8.57 / MAX: 8.75
oneDNN Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU c a b 500 1000 1500 2000 2500 2320.01 2506.07 2555.28 MIN: 2157.6 MIN: 2464.71 MIN: 2505.86 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU c a b 500 1000 1500 2000 2500 2382.55 2520.79 2524.94 MIN: 2326.78 MIN: 2471.71 MIN: 2468.28 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU c a b 500 1000 1500 2000 2500 2127.79 2514.18 2555.43 MIN: 2040.61 MIN: 2466.17 MIN: 2502.86 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
Embree Binary: Pathtracer ISPC - Model: Asian Dragon OpenBenchmarking.org Frames Per Second, More Is Better Embree 4.3 Binary: Pathtracer ISPC - Model: Asian Dragon b a c 3 6 9 12 15 9.5178 9.5152 9.4660 MIN: 9.39 / MAX: 9.87 MIN: 9.42 / MAX: 9.86 MIN: 9.39 / MAX: 9.76
QuantLib Configuration: Multi-Threaded OpenBenchmarking.org MFLOPS, More Is Better QuantLib 1.32 Configuration: Multi-Threaded b a c 6K 12K 18K 24K 30K 28768.7 28605.0 28299.9 1. (CXX) g++ options: -O3 -march=native -fPIE -pie
QuantLib Configuration: Single-Threaded OpenBenchmarking.org MFLOPS, More Is Better QuantLib 1.32 Configuration: Single-Threaded b a c 900 1800 2700 3600 4500 4229.3 4133.0 4074.6 1. (CXX) g++ options: -O3 -march=native -fPIE -pie
oneDNN Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU c b a 3 6 9 12 15 8.70755 9.73785 9.93332 MIN: 7.2 MIN: 8.44 MIN: 8.67 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU c a b 0.3774 0.7548 1.1322 1.5096 1.887 1.29785 1.52901 1.67745 MIN: 1.08 MIN: 1.26 MIN: 1.3 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU c b a 3 6 9 12 15 9.6233 10.4152 10.8682 MIN: 5.21 MIN: 6.45 MIN: 6.35 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU c a b 3 6 9 12 15 8.66246 8.86916 8.99248 MIN: 7.18 MIN: 7.17 MIN: 8.17 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU c b a 0.6204 1.2408 1.8612 2.4816 3.102 2.42519 2.75383 2.75717 MIN: 2.15 MIN: 2.29 MIN: 2.4 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU c b a 0.2882 0.5764 0.8646 1.1528 1.441 1.15590 1.25823 1.28110 MIN: 0.9 MIN: 1.01 MIN: 1.02 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
easyWave Input: e2Asean Grid + BengkuluSept2007 Source - Time: 240 OpenBenchmarking.org Seconds, Fewer Is Better easyWave r34 Input: e2Asean Grid + BengkuluSept2007 Source - Time: 240 b a c 3 6 9 12 15 11.93 11.98 11.98 1. (CXX) g++ options: -O3 -fopenmp
oneDNN Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU c b a 0.8533 1.7066 2.5599 3.4132 4.2665 3.62443 3.75671 3.79236 MIN: 3.4 MIN: 3.37 MIN: 3.42 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU b c a 1.2625 2.525 3.7875 5.05 6.3125 5.57229 5.59833 5.61120 MIN: 5.48 MIN: 5.5 MIN: 5.48 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU c b a 0.501 1.002 1.503 2.004 2.505 2.19780 2.22532 2.22678 MIN: 2.06 MIN: 2.04 MIN: 2.03 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU b a c 4 8 12 16 20 14.94 15.02 15.09 MIN: 14.72 MIN: 14.65 MIN: 14.85 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU a c b 4 8 12 16 20 14.25 14.27 14.58 MIN: 14.01 MIN: 13.71 MIN: 14.19 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU c b a 2 4 6 8 10 6.72684 6.74215 6.76831 MIN: 6.42 MIN: 6.46 MIN: 6.43 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU c a b 2 4 6 8 10 6.36614 7.60937 7.67500 MIN: 4.81 MIN: 6.03 MIN: 6.03 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU c a b 1.1754 2.3508 3.5262 4.7016 5.877 4.42411 5.17232 5.22381 MIN: 3.56 MIN: 3.57 MIN: 3.56 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.3 Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU c a b 0.4524 0.9048 1.3572 1.8096 2.262 1.55587 1.92455 2.01069 MIN: 1.16 MIN: 1.46 MIN: 1.46 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
Phoronix Test Suite v10.8.4