Intel Core i5-12400 testing with a ASRock B660M-HDV (3.02 BIOS) and llvmpipe on Ubuntu 22.04 via the Phoronix Test Suite.
A Processor: Intel Core i5-12400 @ 5.60GHz (6 Cores / 12 Threads), Motherboard: ASRock B660M-HDV (3.02 BIOS), Chipset: Intel Device 7aa7, Memory: 16GB, Disk: 512GB Sabrent, Graphics: llvmpipe, Audio: Realtek ALC897, Network: Intel
OS: Ubuntu 22.04, Kernel: 5.15.0-18-generic (x86_64), Desktop: GNOME Shell 41.3, Display Server: X Server 1.20.14, OpenGL: 4.5 Mesa 21.2.2 (LLVM 12.0.1 256 bits), Vulkan: 1.1.182, Compiler: GCC 11.2.0, File-System: ext4, Screen Resolution: 1920x1080
Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-iOLsLC/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-iOLsLC/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.7Java Notes: OpenJDK Runtime Environment (build 11.0.14.1+1-Ubuntu-0ubuntu1)Python Notes: Python 3.9.12Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Intel Core i5-12400 Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.9Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
B Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.9Java Notes: OpenJDK Runtime Environment (build 11.0.14.1+1-Ubuntu-0ubuntu1)Python Notes: Python 3.10.4Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
C OS: Ubuntu 22.04, Kernel: 5.15.0-18-generic (x86_64), Desktop: GNOME Shell 41.3, Display Server: X Server 1.20.14, OpenGL: 4.5 Mesa 22.0.1 (LLVM 13.0.1 256 bits), Vulkan: 1.2.204, Compiler: GCC 11.2.0, File-System: ext4, Screen Resolution: 1920x1080
core-i5-12400-april Processor Motherboard Chipset Memory Disk Graphics Audio Network OS Kernel Desktop Display Server OpenGL Vulkan Compiler File-System Screen Resolution A Intel Core i5-12400 B C Intel Core i5-12400 @ 5.60GHz (6 Cores / 12 Threads) ASRock B660M-HDV (3.02 BIOS) Intel Device 7aa7 16GB 512GB Sabrent llvmpipe Realtek ALC897 Intel Ubuntu 22.04 5.15.0-18-generic (x86_64) GNOME Shell 41.3 X Server 1.20.14 4.5 Mesa 21.2.2 (LLVM 12.0.1 256 bits) 1.1.182 GCC 11.2.0 ext4 1920x1080 4.5 Mesa 22.0.1 (LLVM 13.0.1 256 bits) 1.2.204 OpenBenchmarking.org Kernel Details - Transparent Huge Pages: madvise Compiler Details - A: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-iOLsLC/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-iOLsLC/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Intel Core i5-12400: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - B: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - C: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details - A: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.7 - Intel Core i5-12400: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.9 - B: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.9 - C: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.9 Java Details - A, B, C: OpenJDK Runtime Environment (build 11.0.14.1+1-Ubuntu-0ubuntu1) Python Details - A: Python 3.9.12 - B: Python 3.10.4 - C: Python 3.10.4 Security Details - itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
core-i5-12400-april onednn: Recurrent Neural Network Training - bf16bf16bf16 - CPU onednn: Deconvolution Batch shapes_3d - u8s8f32 - CPU onednn: Recurrent Neural Network Training - u8s8f32 - CPU onednn: Recurrent Neural Network Training - f32 - CPU onednn: IP Shapes 3D - f32 - CPU onednn: Deconvolution Batch shapes_3d - f32 - CPU onednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU onednn: Deconvolution Batch shapes_1d - u8s8f32 - CPU onednn: IP Shapes 1D - f32 - CPU onednn: Convolution Batch Shapes Auto - u8s8f32 - CPU onednn: Convolution Batch Shapes Auto - f32 - CPU onednn: IP Shapes 1D - u8s8f32 - CPU onednn: Matrix Multiply Batch Shapes Transformer - f32 - CPU onednn: Recurrent Neural Network Inference - f32 - CPU onednn: IP Shapes 3D - u8s8f32 - CPU onednn: Recurrent Neural Network Inference - u8s8f32 - CPU onednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPU onnx: fcn-resnet101-11 - CPU - Parallel aom-av1: Speed 10 Realtime - Bosphorus 4K onnx: super-resolution-10 - CPU - Parallel aom-av1: Speed 10 Realtime - Bosphorus 1080p ospray: particle_volume/scivis/real_time aom-av1: Speed 4 Two-Pass - Bosphorus 4K ospray-studio: 2 - 1080p - 16 - Path Tracer aom-av1: Speed 6 Two-Pass - Bosphorus 4K ospray-studio: 1 - 1080p - 16 - Path Tracer ospray-studio: 1 - 1080p - 32 - Path Tracer dav1d: Summer Nature 1080p ospray-studio: 3 - 1080p - 1 - Path Tracer ospray-studio: 2 - 1080p - 32 - Path Tracer ospray-studio: 3 - 1080p - 32 - Path Tracer aom-av1: Speed 6 Two-Pass - Bosphorus 1080p dav1d: Chimera 1080p 10-bit ospray: particle_volume/pathtracer/real_time dav1d: Summer Nature 4K compress-pbzip2: FreeBSD-13.0-RELEASE-amd64-memstick.img Compression onnx: ArcFace ResNet-100 - CPU - Parallel ospray: gravity_spheres_volume/dim_512/pathtracer/real_time ospray-studio: 3 - 1080p - 16 - Path Tracer build-wasmer: Time To Compile ospray-studio: 2 - 1080p - 1 - Path Tracer java-jmh: Throughput ospray: particle_volume/ao/real_time ospray-studio: 1 - 1080p - 1 - Path Tracer rocksdb: Update Rand ospray: gravity_spheres_volume/dim_512/scivis/real_time dav1d: Chimera 1080p rocksdb: Read While Writing rocksdb: Read Rand Write Rand ospray: gravity_spheres_volume/dim_512/ao/real_time aom-av1: Speed 8 Realtime - Bosphorus 1080p aom-av1: Speed 0 Two-Pass - Bosphorus 1080p onnx: yolov4 - CPU - Parallel build-mplayer: Time To Compile rocksdb: Rand Read aom-av1: Speed 4 Two-Pass - Bosphorus 1080p aom-av1: Speed 6 Realtime - Bosphorus 4K aom-av1: Speed 9 Realtime - Bosphorus 4K aom-av1: Speed 0 Two-Pass - Bosphorus 4K aom-av1: Speed 8 Realtime - Bosphorus 4K aom-av1: Speed 6 Realtime - Bosphorus 1080p onnx: yolov4 - CPU - Standard onnx: GPT-2 - CPU - Parallel onnx: super-resolution-10 - CPU - Standard onnx: fcn-resnet101-11 - CPU - Standard onnx: GPT-2 - CPU - Standard onnx: bertsquad-12 - CPU - Standard onnx: bertsquad-12 - CPU - Parallel onnx: ArcFace ResNet-100 - CPU - Standard avifenc: 2 avifenc: 10, Lossless avifenc: 0 avifenc: 6, Lossless avifenc: 6 onednn: Deconvolution Batch shapes_1d - f32 - CPU aom-av1: Speed 9 Realtime - Bosphorus 1080p A Intel Core i5-12400 B C 36805.2 17.2152 34757.4 29463.3 64.3737 46.6701 53.7011 46.3983 113.724 50.7024 47.6937 28.3268 71.8268 34134.5 27.9172 28488.2 24467.5 44 61.38 2845 133.21 12.3203 4.17 57299 8 56352 115145 710.02 4189 116689 136929 24.11 481.75 163.591 172.71 10.955 853 2.47222 68568 85.485 3471 20972409216.502 12.9688 3398 679780 1.79913 561.82 1467166 1497554 1.83804 107.06 0.34 277 42.432 45697206 9.58 13.91 63.16 0.13 42.47 11.67 292 5745 2822 42 6878 439 427 869 162.853 124.6 85.16 6.2 189.74 14.918 12.207 4099.42 1.98156 4102.34 4095.96 9.90106 7.69890 1.04716 1.42641 3.74052 13.7565 14.3318 0.957803 2.68966 2264.90 2.10894 2264.98 2271.69 56 72.38 3247 150.33 13.8893 4.70 50897 8.99 50151 102548 795.53 3740 104124 122366 26.97 536.82 182.112 191.89 9.890 944 2.73146 62054 77.914 3165 23002668637.36 14.2127 3114 742880 1.96577 612.15 1558793 1631656 2.00038 116.51 0.37 301 39.113 49594691 10.39 15.08 68.14 0.14 45.65 12.36 309 6076 2940 44 7205 458 443 884 85.968 6.247 191.075 14.975 12.213 9.06388 149.25 4098.3 1.99365 4101.91 4086.37 9.84082 7.81726 1.05341 1.42738 3.73962 13.8003 14.3255 0.959336 2.69025 2264.46 2.09714 2261.84 2270.92 57 72.41 3257 149.66 13.8804 4.70 50954 8.92 50418 102588 796.46 3735 104044 122245 26.95 536.42 182.037 192.02 10.031 944 2.73399 62054 77.817 3163 22882301161.56 14.1561 3108 743135 1.95220 612.41 1598728 1621120 1.98550 114.84 0.37 301 39.062 49606865 10.37 15.02 68.39 0.14 45.68 12.41 309 6049 2963 44 7194 458 438 884 85.939 6.257 191.476 14.999 12.193 8.90115 142.90 OpenBenchmarking.org
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU A B C 8K 16K 24K 32K 40K SE +/- 7.00, N = 3 SE +/- 6.25, N = 3 36805.20 4099.42 4098.30 MIN: 16567 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU A C B 4 8 12 16 20 SE +/- 0.01869, N = 3 SE +/- 0.00178, N = 3 17.21520 1.99365 1.98156 MIN: 1.97 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU A B C 7K 14K 21K 28K 35K SE +/- 1.65, N = 3 SE +/- 3.15, N = 3 34757.40 4102.34 4101.91 MIN: 18514.7 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU A B C 6K 12K 18K 24K 30K SE +/- 3.35, N = 3 SE +/- 12.29, N = 3 29463.30 4095.96 4086.37 MIN: 14417.6 MIN: 4045.08 MIN: 4011.27 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU A B C 14 28 42 56 70 SE +/- 0.01269, N = 3 SE +/- 0.00524, N = 3 64.37370 9.90106 9.84082 MIN: 14.76 MIN: 9.49 MIN: 9.41 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU A C B 11 22 33 44 55 SE +/- 0.10400, N = 3 SE +/- 0.00768, N = 3 46.67010 7.81726 7.69890 MIN: 16.15 MIN: 7.6 MIN: 7.57 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU A C B 12 24 36 48 60 SE +/- 0.00112, N = 3 SE +/- 0.00153, N = 3 53.70110 1.05341 1.04716 MIN: 1.03 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU A C B 11 22 33 44 55 SE +/- 0.00246, N = 3 SE +/- 0.00232, N = 3 46.39830 1.42738 1.42641 MIN: 1.4 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU A B C 30 60 90 120 150 SE +/- 0.00457, N = 3 SE +/- 0.00589, N = 3 113.72400 3.74052 3.73962 MIN: 51.18 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU A C B 11 22 33 44 55 SE +/- 0.03, N = 3 SE +/- 0.02, N = 3 50.70 13.80 13.76 MIN: 19.83 MIN: 13.2 MIN: 13.16 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU A B C 11 22 33 44 55 SE +/- 0.01, N = 3 SE +/- 0.01, N = 3 47.69 14.33 14.33 MIN: 17.03 MIN: 14.19 MIN: 14.16 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU A C B 7 14 21 28 35 SE +/- 0.001531, N = 3 SE +/- 0.001376, N = 3 28.326800 0.959336 0.957803 MIN: 0.92 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU A C B 16 32 48 64 80 SE +/- 0.00320, N = 3 SE +/- 0.00256, N = 3 71.82680 2.69025 2.68966 MIN: 3.12 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU A B C 7K 14K 21K 28K 35K SE +/- 0.80, N = 3 SE +/- 3.21, N = 3 34134.50 2264.90 2264.46 MIN: 18540.8 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU A B C 7 14 21 28 35 SE +/- 0.00109, N = 3 SE +/- 0.00196, N = 3 27.91720 2.10894 2.09714 MIN: 2.08 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU A B C 6K 12K 18K 24K 30K SE +/- 1.35, N = 3 SE +/- 5.98, N = 3 28488.20 2264.98 2261.84 MIN: 12419 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU A B C 5K 10K 15K 20K 25K SE +/- 4.51, N = 3 SE +/- 3.05, N = 3 24467.50 2271.69 2270.92 MIN: 9924.5 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
ONNX Runtime ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: fcn-resnet101-11 - Device: CPU - Executor: Parallel A B C 13 26 39 52 65 SE +/- 0.44, N = 3 SE +/- 0.17, N = 3 44 56 57 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
AOM AV1 OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 10 Realtime - Input: Bosphorus 4K A B C 16 32 48 64 80 SE +/- 0.14, N = 3 SE +/- 0.05, N = 3 61.38 72.38 72.41 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
ONNX Runtime ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: super-resolution-10 - Device: CPU - Executor: Parallel A B C 700 1400 2100 2800 3500 SE +/- 4.73, N = 3 SE +/- 11.90, N = 3 2845 3247 3257 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
AOM AV1 OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 10 Realtime - Input: Bosphorus 1080p A C B 30 60 90 120 150 SE +/- 1.27, N = 3 SE +/- 0.90, N = 3 133.21 149.66 150.33 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OSPray OpenBenchmarking.org Items Per Second, More Is Better OSPray 2.9 Benchmark: particle_volume/scivis/real_time A C B 4 8 12 16 20 SE +/- 0.02, N = 3 SE +/- 0.01, N = 3 12.32 13.88 13.89
AOM AV1 OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 4K A B C 1.0575 2.115 3.1725 4.23 5.2875 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 4.17 4.70 4.70 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OSPray Studio Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer A C B 12K 24K 36K 48K 60K SE +/- 27.83, N = 3 SE +/- 44.19, N = 3 57299 50954 50897 1. (CXX) g++ options: -O3 -lm -ldl
AOM AV1 OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 4K A C B 3 6 9 12 15 SE +/- 0.07, N = 3 SE +/- 0.00, N = 3 8.00 8.92 8.99 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OSPray Studio Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer A C B 12K 24K 36K 48K 60K SE +/- 272.17, N = 3 SE +/- 21.38, N = 3 56352 50418 50151 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer A C B 20K 40K 60K 80K 100K SE +/- 84.91, N = 3 SE +/- 146.65, N = 3 115145 102588 102548 1. (CXX) g++ options: -O3 -lm -ldl
OSPray Studio Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer A B C 900 1800 2700 3600 4500 SE +/- 3.53, N = 3 SE +/- 2.85, N = 3 4189 3740 3735 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer A B C 20K 40K 60K 80K 100K SE +/- 167.02, N = 3 SE +/- 63.66, N = 3 116689 104124 104044 1. (CXX) g++ options: -O3 -lm -ldl
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path Tracer A B C 30K 60K 90K 120K 150K SE +/- 170.37, N = 3 SE +/- 103.73, N = 3 136929 122366 122245 1. (CXX) g++ options: -O3 -lm -ldl
AOM AV1 OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 1080p A C B 6 12 18 24 30 SE +/- 0.00, N = 3 SE +/- 0.01, N = 3 24.11 26.95 26.97 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OSPray OpenBenchmarking.org Items Per Second, More Is Better OSPray 2.9 Benchmark: particle_volume/pathtracer/real_time A C B 40 80 120 160 200 SE +/- 0.10, N = 3 SE +/- 0.02, N = 3 163.59 182.04 182.11
ONNX Runtime ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel A B C 200 400 600 800 1000 SE +/- 3.33, N = 3 SE +/- 2.96, N = 3 853 944 944 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OSPray OpenBenchmarking.org Items Per Second, More Is Better OSPray 2.9 Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_time A B C 0.6151 1.2302 1.8453 2.4604 3.0755 SE +/- 0.00279, N = 3 SE +/- 0.00158, N = 3 2.47222 2.73146 2.73399
OSPray Studio Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 3 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path Tracer A C B 15K 30K 45K 60K 75K SE +/- 140.67, N = 3 SE +/- 147.67, N = 3 68568 62054 62054 1. (CXX) g++ options: -O3 -lm -ldl
Timed Wasmer Compilation This test times how long it takes to compile Wasmer. Wasmer is written in the Rust programming language and is a WebAssembly runtime implementation that supports WASI and EmScripten. This test profile builds Wasmer with the Cranelift and Singlepast compiler features enabled. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Timed Wasmer Compilation 2.2 Time To Compile A B C 20 40 60 80 100 SE +/- 0.11, N = 3 SE +/- 0.23, N = 3 85.49 77.91 77.82 1. (CC) gcc options: -m64 -ldl -lxkbcommon -lgcc_s -lutil -lrt -lpthread -lm -lc -pie -nodefaultlibs
OSPray Studio Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 2 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer A B C 700 1400 2100 2800 3500 SE +/- 0.58, N = 3 SE +/- 0.33, N = 3 3471 3165 3163 1. (CXX) g++ options: -O3 -lm -ldl
OSPray OpenBenchmarking.org Items Per Second, More Is Better OSPray 2.9 Benchmark: particle_volume/ao/real_time A C B 4 8 12 16 20 SE +/- 0.04, N = 3 SE +/- 0.03, N = 3 12.97 14.16 14.21
OSPray Studio Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better OSPray Studio 0.10 Camera: 1 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path Tracer A B C 700 1400 2100 2800 3500 SE +/- 7.42, N = 3 SE +/- 10.02, N = 3 3398 3114 3108 1. (CXX) g++ options: -O3 -lm -ldl
OSPray OpenBenchmarking.org Items Per Second, More Is Better OSPray 2.9 Benchmark: gravity_spheres_volume/dim_512/scivis/real_time A C B 0.4423 0.8846 1.3269 1.7692 2.2115 SE +/- 0.00260, N = 3 SE +/- 0.00094, N = 3 1.79913 1.95220 1.96577
OpenBenchmarking.org Op/s, More Is Better Facebook RocksDB 7.0.1 Test: Read Random Write Random A C B 300K 600K 900K 1200K 1500K SE +/- 1481.37, N = 3 SE +/- 8702.90, N = 3 1497554 1621120 1631656 1. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread
OSPray OpenBenchmarking.org Items Per Second, More Is Better OSPray 2.9 Benchmark: gravity_spheres_volume/dim_512/ao/real_time A C B 0.4501 0.9002 1.3503 1.8004 2.2505 SE +/- 0.00129, N = 3 SE +/- 0.00584, N = 3 1.83804 1.98550 2.00038
AOM AV1 OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 8 Realtime - Input: Bosphorus 1080p A C B 30 60 90 120 150 SE +/- 0.29, N = 3 SE +/- 0.41, N = 3 107.06 114.84 116.51 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 1080p A B C 0.0833 0.1666 0.2499 0.3332 0.4165 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 0.34 0.37 0.37 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
ONNX Runtime ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: yolov4 - Device: CPU - Executor: Parallel A B C 70 140 210 280 350 SE +/- 0.88, N = 3 SE +/- 1.36, N = 3 277 301 301 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
AOM AV1 OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 1080p A C B 3 6 9 12 15 SE +/- 0.04, N = 3 SE +/- 0.01, N = 3 9.58 10.37 10.39 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 6 Realtime - Input: Bosphorus 4K A C B 4 8 12 16 20 SE +/- 0.07, N = 3 SE +/- 0.13, N = 15 13.91 15.02 15.08 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 9 Realtime - Input: Bosphorus 4K A B C 15 30 45 60 75 SE +/- 0.12, N = 3 SE +/- 0.04, N = 3 63.16 68.14 68.39 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 4K A B C 0.0315 0.063 0.0945 0.126 0.1575 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 0.13 0.14 0.14 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 8 Realtime - Input: Bosphorus 4K A B C 10 20 30 40 50 SE +/- 0.02, N = 3 SE +/- 0.07, N = 3 42.47 45.65 45.68 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 6 Realtime - Input: Bosphorus 1080p A B C 3 6 9 12 15 SE +/- 0.04, N = 3 SE +/- 0.10, N = 8 11.67 12.36 12.41 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
ONNX Runtime ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: yolov4 - Device: CPU - Executor: Standard A B C 70 140 210 280 350 SE +/- 0.17, N = 3 SE +/- 0.17, N = 3 292 309 309 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: GPT-2 - Device: CPU - Executor: Parallel A C B 1300 2600 3900 5200 6500 SE +/- 8.66, N = 3 SE +/- 5.13, N = 3 5745 6049 6076 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: super-resolution-10 - Device: CPU - Executor: Standard A B C 600 1200 1800 2400 3000 SE +/- 12.08, N = 3 SE +/- 0.67, N = 3 2822 2940 2963 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: fcn-resnet101-11 - Device: CPU - Executor: Standard A B C 10 20 30 40 50 SE +/- 0.00, N = 3 SE +/- 0.00, N = 3 42 44 44 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: GPT-2 - Device: CPU - Executor: Standard A C B 1500 3000 4500 6000 7500 SE +/- 2.52, N = 3 SE +/- 6.71, N = 3 6878 7194 7205 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: bertsquad-12 - Device: CPU - Executor: Standard A B C 100 200 300 400 500 SE +/- 0.17, N = 3 SE +/- 0.17, N = 3 439 458 458 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: bertsquad-12 - Device: CPU - Executor: Parallel A C B 100 200 300 400 500 SE +/- 3.68, N = 12 SE +/- 4.83, N = 12 427 438 443 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Inferences Per Minute, More Is Better ONNX Runtime 1.11 Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard A B C 200 400 600 800 1000 SE +/- 0.00, N = 3 SE +/- 0.17, N = 3 869 884 884 1. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt
OpenBenchmarking.org Seconds, Fewer Is Better libavif avifenc 0.10 Encoder Speed: 10, Lossless C B Intel Core i5-12400 2 4 6 8 10 SE +/- 0.029, N = 3 SE +/- 0.026, N = 3 6.257 6.247 6.200 1. (CXX) g++ options: -O3 -fPIC -lm
OpenBenchmarking.org Seconds, Fewer Is Better libavif avifenc 0.10 Encoder Speed: 0 C B Intel Core i5-12400 40 80 120 160 200 SE +/- 0.20, N = 3 SE +/- 0.09, N = 3 191.48 191.08 189.74 1. (CXX) g++ options: -O3 -fPIC -lm
OpenBenchmarking.org Seconds, Fewer Is Better libavif avifenc 0.10 Encoder Speed: 6, Lossless C B Intel Core i5-12400 4 8 12 16 20 SE +/- 0.03, N = 3 SE +/- 0.03, N = 3 15.00 14.98 14.92 1. (CXX) g++ options: -O3 -fPIC -lm
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.
Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU
A: The test run did not produce a result.
B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.
C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.
Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU
A: The test run did not produce a result.
B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.
C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.
Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU
A: The test run did not produce a result.
B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.
C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.
Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU
A: The test run did not produce a result.
B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.
C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU A B C 40 80 120 160 200 SE +/- 0.17309, N = 15 SE +/- 0.13280, N = 15 162.85300 9.06388 8.90115 MIN: 6.38 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl
Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU
A: The test run did not produce a result.
B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.
C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.
Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU
A: The test run did not produce a result.
B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.
C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.
AOM AV1 OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.3 Encoder Mode: Speed 9 Realtime - Input: Bosphorus 1080p A C B 30 60 90 120 150 SE +/- 0.31, N = 3 SE +/- 2.69, N = 15 124.60 142.90 149.25 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm
A Processor: Intel Core i5-12400 @ 5.60GHz (6 Cores / 12 Threads), Motherboard: ASRock B660M-HDV (3.02 BIOS), Chipset: Intel Device 7aa7, Memory: 16GB, Disk: 512GB Sabrent, Graphics: llvmpipe, Audio: Realtek ALC897, Network: Intel
OS: Ubuntu 22.04, Kernel: 5.15.0-18-generic (x86_64), Desktop: GNOME Shell 41.3, Display Server: X Server 1.20.14, OpenGL: 4.5 Mesa 21.2.2 (LLVM 12.0.1 256 bits), Vulkan: 1.1.182, Compiler: GCC 11.2.0, File-System: ext4, Screen Resolution: 1920x1080
Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-iOLsLC/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-iOLsLC/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.7Java Notes: OpenJDK Runtime Environment (build 11.0.14.1+1-Ubuntu-0ubuntu1)Python Notes: Python 3.9.12Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 6 April 2022 15:24 by user phoronix.
Intel Core i5-12400 Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.9Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 6 April 2022 19:43 by user phoronix.
B Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.9Java Notes: OpenJDK Runtime Environment (build 11.0.14.1+1-Ubuntu-0ubuntu1)Python Notes: Python 3.10.4Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 6 April 2022 20:29 by user phoronix.
C Processor: Intel Core i5-12400 @ 5.60GHz (6 Cores / 12 Threads), Motherboard: ASRock B660M-HDV (3.02 BIOS), Chipset: Intel Device 7aa7, Memory: 16GB, Disk: 512GB Sabrent, Graphics: llvmpipe, Audio: Realtek ALC897, Network: Intel
OS: Ubuntu 22.04, Kernel: 5.15.0-18-generic (x86_64), Desktop: GNOME Shell 41.3, Display Server: X Server 1.20.14, OpenGL: 4.5 Mesa 22.0.1 (LLVM 13.0.1 256 bits), Vulkan: 1.2.204, Compiler: GCC 11.2.0, File-System: ext4, Screen Resolution: 1920x1080
Kernel Notes: Transparent Huge Pages: madviseCompiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.9Java Notes: OpenJDK Runtime Environment (build 11.0.14.1+1-Ubuntu-0ubuntu1)Python Notes: Python 3.10.4Security Notes: itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected
Testing initiated at 7 April 2022 05:56 by user phoronix.