new tests icelake Intel Core i7-1065G7 testing with a Dell 06CDVY (1.0.9 BIOS) and Intel Iris Plus G7 3GB on Ubuntu 20.04 via the Phoronix Test Suite.
Compare your own system(s) to this result file with the
Phoronix Test Suite by running the command:
phoronix-test-suite benchmark 2006245-NE-NEWTESTSI85 Intel Killer Wi-Fi 6 AX1650i 160MHz Processor: Intel Core i7-1065G7 @ 3.90GHz (4 Cores / 8 Threads), Motherboard: Dell 06CDVY (1.0.9 BIOS), Chipset: Intel Device 34ef, Memory: 16GB, Disk: KBG40ZPZ512G NVMe TOSHIBA 512GB, Graphics: Intel Iris Plus G7 3GB (1100MHz), Audio: Realtek ALC289, Network: Intel Killer Wi-Fi 6 AX1650i 160MHz
OS: Ubuntu 20.04, Kernel: 5.4.0-33-generic (x86_64), Desktop: GNOME Shell 3.36.2, Display Server: X Server 1.20.8, Display Driver: modesetting 1.20.8, OpenGL: 4.6 Mesa 20.0.4, Vulkan: 1.2.131, Compiler: GCC 9.3.0, File-System: ext4, Screen Resolution: 1920x1200
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave - CPU Microcode: 0x78Security Notes: itlb_multihit: KVM: Mitigation of Split huge pages + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + tsx_async_abort: Not affected
new tests icelake OpenBenchmarking.org Phoronix Test Suite Intel Core i7-1065G7 @ 3.90GHz (4 Cores / 8 Threads) Dell 06CDVY (1.0.9 BIOS) Intel Device 34ef 16GB KBG40ZPZ512G NVMe TOSHIBA 512GB Intel Iris Plus G7 3GB (1100MHz) Realtek ALC289 Intel Killer Wi-Fi 6 AX1650i 160MHz Ubuntu 20.04 5.4.0-33-generic (x86_64) GNOME Shell 3.36.2 X Server 1.20.8 modesetting 1.20.8 4.6 Mesa 20.0.4 1.2.131 GCC 9.3.0 ext4 1920x1200 Processor Motherboard Chipset Memory Disk Graphics Audio Network OS Kernel Desktop Display Server Display Driver OpenGL Vulkan Compiler File-System Screen Resolution New Tests Icelake Benchmarks System Logs - --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: intel_pstate powersave - CPU Microcode: 0x78 - itlb_multihit: KVM: Mitigation of Split huge pages + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + tsx_async_abort: Not affected
new tests icelake daphne: OpenMP - NDT Mapping daphne: OpenMP - Points2Image daphne: OpenMP - Euclidean Cluster rodinia: OpenMP LavaMD rodinia: OpenMP Myocyte rodinia: OpenMP HotSpot3D rodinia: OpenMP Leukocyte rodinia: OpenMP CFD Solver rodinia: OpenMP Streamcluster onednn: IP Batch 1D - f32 - CPU onednn: IP Batch All - f32 - CPU onednn: IP Batch 1D - u8s8f32 - CPU onednn: IP Batch All - u8s8f32 - CPU onednn: IP Batch 1D - bf16bf16bf16 - CPU onednn: IP Batch All - bf16bf16bf16 - CPU onednn: Convolution Batch Shapes Auto - f32 - CPU onednn: Deconvolution Batch deconv_1d - f32 - CPU onednn: Deconvolution Batch deconv_3d - f32 - CPU onednn: Convolution Batch Shapes Auto - u8s8f32 - CPU onednn: Deconvolution Batch deconv_1d - u8s8f32 - CPU onednn: Deconvolution Batch deconv_3d - u8s8f32 - CPU onednn: Recurrent Neural Network Training - f32 - CPU onednn: Recurrent Neural Network Inference - f32 - CPU onednn: Convolution Batch Shapes Auto - bf16bf16bf16 - CPU onednn: Deconvolution Batch deconv_1d - bf16bf16bf16 - CPU onednn: Deconvolution Batch deconv_3d - bf16bf16bf16 - CPU onednn: Matrix Multiply Batch Shapes Transformer - f32 - CPU onednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU onednn: Matrix Multiply Batch Shapes Transformer - bf16bf16bf16 - CPU wireguard: build-linux-kernel: Time To Compile Intel Killer Wi-Fi 6 AX1650i 160MHz 854.67 24625.799371187 743.25 846.140 3.519 112.181 329.250 54.380 28.133 14.9091 176.548 3.22710 42.6970 33.0418 448.836 15.5321 15.5462 13.5460 11.5618 3.92116 3.51752 867.418 230.671 63.7207 72.7579 50.3586 5.16681 2.61865 15.7084 290.312 284.797 OpenBenchmarking.org
Darmstadt Automotive Parallel Heterogeneous Suite DAPHNE is the Darmstadt Automotive Parallel HeterogeNEous Benchmark Suite with OpenCL / CUDA / OpenMP test cases for these automotive benchmarks for evaluating programming models in context to vehicle autonomous driving capabilities. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Test Cases Per Minute, More Is Better Darmstadt Automotive Parallel Heterogeneous Suite Backend: OpenMP - Kernel: NDT Mapping Intel Killer Wi-Fi 6 AX1650i 160MHz 200 400 600 800 1000 SE +/- 4.31, N = 3 854.67 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp
Rodinia Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes select OpenCL, NVIDIA CUDA and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP LavaMD Intel Killer Wi-Fi 6 AX1650i 160MHz 200 400 600 800 1000 SE +/- 10.52, N = 3 846.14 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP Myocyte Intel Killer Wi-Fi 6 AX1650i 160MHz 0.7918 1.5836 2.3754 3.1672 3.959 SE +/- 0.024, N = 3 3.519 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP HotSpot3D Intel Killer Wi-Fi 6 AX1650i 160MHz 30 60 90 120 150 SE +/- 0.78, N = 3 112.18 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP Leukocyte Intel Killer Wi-Fi 6 AX1650i 160MHz 70 140 210 280 350 SE +/- 0.87, N = 3 329.25 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP CFD Solver Intel Killer Wi-Fi 6 AX1650i 160MHz 12 24 36 48 60 SE +/- 0.42, N = 3 54.38 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP Streamcluster Intel Killer Wi-Fi 6 AX1650i 160MHz 7 14 21 28 35 SE +/- 0.46, N = 3 28.13 1. (CXX) g++ options: -O2 -lOpenCL
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch 1D - Data Type: f32 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 4 8 12 16 20 SE +/- 0.35, N = 12 14.91 MIN: 8.92 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch All - Data Type: f32 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 40 80 120 160 200 SE +/- 0.26, N = 3 176.55 MIN: 162.61 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch 1D - Data Type: u8s8f32 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 0.7261 1.4522 2.1783 2.9044 3.6305 SE +/- 0.06284, N = 12 3.22710 MIN: 2 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch All - Data Type: u8s8f32 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 10 20 30 40 50 SE +/- 0.04, N = 3 42.70 MIN: 36.76 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch 1D - Data Type: bf16bf16bf16 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 8 16 24 32 40 SE +/- 0.74, N = 12 33.04 MIN: 22.28 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch All - Data Type: bf16bf16bf16 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 100 200 300 400 500 SE +/- 0.61, N = 3 448.84 MIN: 441.5 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 4 8 12 16 20 SE +/- 0.15, N = 12 15.53 MIN: 12.21 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_1d - Data Type: f32 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 4 8 12 16 20 SE +/- 0.23, N = 3 15.55 MIN: 13.42 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_3d - Data Type: f32 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 3 6 9 12 15 SE +/- 0.17, N = 15 13.55 MIN: 11.33 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 3 6 9 12 15 SE +/- 0.12, N = 3 11.56 MIN: 11.18 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 0.8823 1.7646 2.6469 3.5292 4.4115 SE +/- 0.05440, N = 3 3.92116 MIN: 3.05 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 0.7914 1.5828 2.3742 3.1656 3.957 SE +/- 0.05056, N = 15 3.51752 MIN: 2.76 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 200 400 600 800 1000 SE +/- 0.19, N = 3 867.42 MIN: 836.59 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 50 100 150 200 250 SE +/- 1.91, N = 3 230.67 MIN: 214.7 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 14 28 42 56 70 SE +/- 1.69, N = 15 63.72 MIN: 43.27 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_1d - Data Type: bf16bf16bf16 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 16 32 48 64 80 SE +/- 0.85, N = 3 72.76 MIN: 61.5 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_3d - Data Type: bf16bf16bf16 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 11 22 33 44 55 SE +/- 0.65, N = 15 50.36 MIN: 43.86 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 1.1625 2.325 3.4875 4.65 5.8125 SE +/- 0.07701, N = 12 5.16681 MIN: 3.49 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 0.5892 1.1784 1.7676 2.3568 2.946 SE +/- 0.05518, N = 12 2.61865 MIN: 1.61 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU Intel Killer Wi-Fi 6 AX1650i 160MHz 4 8 12 16 20 SE +/- 0.03, N = 3 15.71 MIN: 14.58 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
WireGuard + Linux Networking Stack Stress Test This is a benchmark of the WireGuard secure VPN tunnel and Linux networking stack stress test. The test runs on the local host but does require root permissions to run. The way it works is it creates three namespaces. ns0 has a loopback device. ns1 and ns2 each have wireguard devices. Those two wireguard devices send traffic through the loopback device of ns0. The end result of this is that tests wind up testing encryption and decryption at the same time -- a pretty CPU and scheduler-heavy workflow. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better WireGuard + Linux Networking Stack Stress Test Intel Killer Wi-Fi 6 AX1650i 160MHz 60 120 180 240 300 SE +/- 1.30, N = 3 290.31
Intel Killer Wi-Fi 6 AX1650i 160MHz Processor: Intel Core i7-1065G7 @ 3.90GHz (4 Cores / 8 Threads), Motherboard: Dell 06CDVY (1.0.9 BIOS), Chipset: Intel Device 34ef, Memory: 16GB, Disk: KBG40ZPZ512G NVMe TOSHIBA 512GB, Graphics: Intel Iris Plus G7 3GB (1100MHz), Audio: Realtek ALC289, Network: Intel Killer Wi-Fi 6 AX1650i 160MHz
OS: Ubuntu 20.04, Kernel: 5.4.0-33-generic (x86_64), Desktop: GNOME Shell 3.36.2, Display Server: X Server 1.20.8, Display Driver: modesetting 1.20.8, OpenGL: 4.6 Mesa 20.0.4, Vulkan: 1.2.131, Compiler: GCC 9.3.0, File-System: ext4, Screen Resolution: 1920x1200
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave - CPU Microcode: 0x78Security Notes: itlb_multihit: KVM: Mitigation of Split huge pages + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + tsx_async_abort: Not affected
Testing initiated at 23 June 2020 19:29 by user phoronix.