Intel Core i7-5600U testing with a LENOVO 20BSCTO1WW (N14ET49W 1.27 BIOS) and Intel HD 5500 3GB on Ubuntu 20.04 via the Phoronix Test Suite.
Intel I218-LM - Intel Core i7-5600U Processor: Intel Core i7-5600U @ 3.20GHz (2 Cores / 4 Threads), Motherboard: LENOVO 20BSCTO1WW (N14ET49W 1.27 BIOS), Chipset: Intel Broadwell-U-OPI, Memory: 8GB, Disk: 128GB SAMSUNG MZNTE128, Graphics: Intel HD 5500 3GB (950MHz), Audio: Intel Broadwell-U Audio, Network: Intel I218-LM + Intel 7265
OS: Ubuntu 20.04, Kernel: 5.4.0-40-generic (x86_64), Desktop: GNOME Shell 3.36.1, Display Server: X Server 1.20.8, Display Driver: modesetting 1.20.8, OpenGL: 4.6 Mesa 20.0.4, OpenCL: OpenCL 2.1, Compiler: GCC 9.3.0, File-System: ext4, Screen Resolution: 1920x1080
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave - CPU Microcode: 0x2fPython Notes: Python 3.8.2Security Notes: itlb_multihit: KVM: Mitigation of Split huge pages + l1tf: Mitigation of PTE Inversion; VMX: conditional cache flushes SMT vulnerable + mds: Mitigation of Clear buffers; SMT vulnerable + meltdown: Mitigation of PTI + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + srbds: Mitigation of Microcode + tsx_async_abort: Mitigation of Clear buffers; SMT vulnerable
Linux Laptop OpenBenchmarking.org Phoronix Test Suite Intel Core i7-5600U @ 3.20GHz (2 Cores / 4 Threads) LENOVO 20BSCTO1WW (N14ET49W 1.27 BIOS) Intel Broadwell-U-OPI 8GB 128GB SAMSUNG MZNTE128 Intel HD 5500 3GB (950MHz) Intel Broadwell-U Audio Intel I218-LM + Intel 7265 Ubuntu 20.04 5.4.0-40-generic (x86_64) GNOME Shell 3.36.1 X Server 1.20.8 modesetting 1.20.8 4.6 Mesa 20.0.4 OpenCL 2.1 GCC 9.3.0 ext4 1920x1080 Processor Motherboard Chipset Memory Disk Graphics Audio Network OS Kernel Desktop Display Server Display Driver OpenGL OpenCL Compiler File-System Screen Resolution Linux Laptop Benchmarks System Logs - --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: intel_pstate powersave - CPU Microcode: 0x2f - Python 3.8.2 - itlb_multihit: KVM: Mitigation of Split huge pages + l1tf: Mitigation of PTE Inversion; VMX: conditional cache flushes SMT vulnerable + mds: Mitigation of Clear buffers; SMT vulnerable + meltdown: Mitigation of PTI + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + srbds: Mitigation of Microcode + tsx_async_abort: Mitigation of Clear buffers; SMT vulnerable
Linux Laptop brl-cad: VGR Performance Metric tesseract-ocr: Time To OCR 7 Images pyperformance: pickle_pure_python pyperformance: django_template pyperformance: python_startup pyperformance: regex_compile pyperformance: crypto_pyaes pyperformance: json_loads pyperformance: raytrace pyperformance: pathlib pyperformance: nbody pyperformance: float pyperformance: chaos pyperformance: 2to3 pyperformance: go v-ray: CPU octave-benchmark: ocrmypdf: Processing 60 Page PDF Document hugin: Panorama Photo Assistant + Stitching Time daphne: OpenMP - Euclidean Cluster daphne: OpenMP - Points2Image daphne: OpenMP - NDT Mapping montage: Mosaic of M17, K band, 1.5 deg x 1.5 deg build-apache: Time To Compile avifenc: 10 avifenc: 8 avifenc: 2 avifenc: 0 luxcorerender: Rainbow Colors and Prism luxcorerender: DLSC aom-av1: Speed 8 Realtime aom-av1: Speed 6 Two-Pass aom-av1: Speed 6 Realtime aom-av1: Speed 4 Two-Pass aom-av1: Speed 0 Two-Pass dav1d: Chimera 1080p 10-bit dav1d: Summer Nature 1080p dav1d: Summer Nature 4K dav1d: Chimera 1080p onednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU onednn: Matrix Multiply Batch Shapes Transformer - f32 - CPU onednn: Recurrent Neural Network Inference - f32 - CPU onednn: Recurrent Neural Network Training - f32 - CPU onednn: Deconvolution Batch deconv_3d - u8s8f32 - CPU onednn: Deconvolution Batch deconv_1d - u8s8f32 - CPU onednn: Convolution Batch Shapes Auto - u8s8f32 - CPU onednn: Deconvolution Batch deconv_3d - f32 - CPU onednn: Deconvolution Batch deconv_1d - f32 - CPU onednn: Convolution Batch Shapes Auto - f32 - CPU onednn: IP Batch 1D - u8s8f32 - CPU onednn: IP Batch All - f32 - CPU onednn: IP Batch 1D - f32 - CPU compress-zstd: 19 compress-zstd: 3 rodinia: OpenMP Streamcluster rodinia: OpenMP CFD Solver rodinia: OpenMP Leukocyte rodinia: OpenMP HotSpot3D rodinia: OpenCL Myocyte rodinia: OpenMP LavaMD wireguard: onednn: IP Batch All - u8s8f32 - CPU Intel I218-LM - Intel Core i7-5600U 14650 35.174 571 64.8 11.7 225 139 31.2 603 26.8 160 152 139 433 314 2282 10.465 130.368 127.554 496.18 15931.136081342 495.12 109.626 66.183 13.024 14.738 278.495 475.434 0.30 0.27 22.80 1.62 6.59 1.01 0.11 29.77 131.39 35.06 139.86 14.1053 8.70216 659.227 1504.45 25.9140 29.2481 26.3471 33.9677 25.2460 28.1967 13.2751 298.140 21.1719 11.8 1310.3 40.622 101.297 625.912 207.518 110.075 1768.173 385.051 195.808 OpenBenchmarking.org
BRL-CAD BRL-CAD 7.28.0 is a cross-platform, open-source solid modeling system with built-in benchmark mode. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org VGR Performance Metric, More Is Better BRL-CAD 7.30.8 VGR Performance Metric Intel I218-LM - Intel Core i7-5600U 3K 6K 9K 12K 15K 14650 1. (CXX) g++ options: -std=c++11 -pipe -fno-strict-aliasing -fno-common -fexceptions -ftemplate-depth-128 -m64 -ggdb3 -O3 -fipa-pta -fstrength-reduce -finline-functions -flto -pedantic -rdynamic -lSM -lICE -lXi -lGLU -lGL -lGLdispatch -lX11 -lXext -lXrender -lpthread -ldl -luuid -lm
OCRMyPDF OCRMyPDF is an optical character recognition (OCR) text layer to scanned PDF files, producing new PDFs with the text now selectable/searchable/copy-paste capable. OCRMyPDF leverages the Tesseract OCR engine and is written in Python. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better OCRMyPDF 9.6.0+dfsg Processing 60 Page PDF Document Intel I218-LM - Intel Core i7-5600U 30 60 90 120 150 SE +/- 1.50, N = 3 130.37
Hugin Hugin is an open-source, cross-platform panorama photo stitcher software package. This test profile times how long it takes to run the assistant and panorama photo stitching on a set of images. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Hugin Panorama Photo Assistant + Stitching Time Intel I218-LM - Intel Core i7-5600U 30 60 90 120 150 SE +/- 0.38, N = 3 127.55
Darmstadt Automotive Parallel Heterogeneous Suite DAPHNE is the Darmstadt Automotive Parallel HeterogeNEous Benchmark Suite with OpenCL / CUDA / OpenMP test cases for these automotive benchmarks for evaluating programming models in context to vehicle autonomous driving capabilities. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Test Cases Per Minute, More Is Better Darmstadt Automotive Parallel Heterogeneous Suite Backend: OpenMP - Kernel: Euclidean Cluster Intel I218-LM - Intel Core i7-5600U 110 220 330 440 550 SE +/- 0.13, N = 3 496.18 1. (CXX) g++ options: -O3 -std=c++11 -fopenmp
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 2.0 Encoder Mode: Speed 6 Two-Pass Intel I218-LM - Intel Core i7-5600U 0.3645 0.729 1.0935 1.458 1.8225 SE +/- 0.01, N = 3 1.62 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 2.0 Encoder Mode: Speed 6 Realtime Intel I218-LM - Intel Core i7-5600U 2 4 6 8 10 SE +/- 0.04, N = 3 6.59 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 2.0 Encoder Mode: Speed 4 Two-Pass Intel I218-LM - Intel Core i7-5600U 0.2273 0.4546 0.6819 0.9092 1.1365 SE +/- 0.01, N = 3 1.01 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 2.0 Encoder Mode: Speed 0 Two-Pass Intel I218-LM - Intel Core i7-5600U 0.0248 0.0496 0.0744 0.0992 0.124 SE +/- 0.00, N = 15 0.11 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.org FPS, More Is Better dav1d 0.7.0 Video Input: Summer Nature 1080p Intel I218-LM - Intel Core i7-5600U 30 60 90 120 150 SE +/- 0.04, N = 3 131.39 MIN: 120.43 / MAX: 146.6 1. (CC) gcc options: -pthread -ldl -lm
OpenBenchmarking.org FPS, More Is Better dav1d 0.7.0 Video Input: Summer Nature 4K Intel I218-LM - Intel Core i7-5600U 8 16 24 32 40 SE +/- 0.05, N = 3 35.06 MIN: 31.36 / MAX: 43.64 1. (CC) gcc options: -pthread -ldl -lm
OpenBenchmarking.org FPS, More Is Better dav1d 0.7.0 Video Input: Chimera 1080p Intel I218-LM - Intel Core i7-5600U 30 60 90 120 150 SE +/- 0.54, N = 3 139.86 MIN: 97.03 / MAX: 244.15 1. (CC) gcc options: -pthread -ldl -lm
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU Intel I218-LM - Intel Core i7-5600U 4 8 12 16 20 SE +/- 0.14, N = 3 14.11 MIN: 13.48 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU Intel I218-LM - Intel Core i7-5600U 2 4 6 8 10 SE +/- 0.05510, N = 3 8.70216 MIN: 8.44 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU Intel I218-LM - Intel Core i7-5600U 140 280 420 560 700 SE +/- 8.45, N = 4 659.23 MIN: 629.31 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU Intel I218-LM - Intel Core i7-5600U 300 600 900 1200 1500 SE +/- 24.80, N = 3 1504.45 MIN: 1404.66 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32 - Engine: CPU Intel I218-LM - Intel Core i7-5600U 6 12 18 24 30 SE +/- 0.03, N = 3 25.91 MIN: 25.83 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32 - Engine: CPU Intel I218-LM - Intel Core i7-5600U 7 14 21 28 35 SE +/- 0.47, N = 3 29.25 MIN: 28.07 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU Intel I218-LM - Intel Core i7-5600U 6 12 18 24 30 SE +/- 0.02, N = 3 26.35 MIN: 25.95 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_3d - Data Type: f32 - Engine: CPU Intel I218-LM - Intel Core i7-5600U 8 16 24 32 40 SE +/- 0.03, N = 3 33.97 MIN: 33.17 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_1d - Data Type: f32 - Engine: CPU Intel I218-LM - Intel Core i7-5600U 6 12 18 24 30 SE +/- 0.36, N = 3 25.25 MIN: 23.96 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU Intel I218-LM - Intel Core i7-5600U 7 14 21 28 35 SE +/- 0.06, N = 3 28.20 MIN: 27.98 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch 1D - Data Type: u8s8f32 - Engine: CPU Intel I218-LM - Intel Core i7-5600U 3 6 9 12 15 SE +/- 0.08, N = 3 13.28 MIN: 12.97 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch All - Data Type: f32 - Engine: CPU Intel I218-LM - Intel Core i7-5600U 60 120 180 240 300 SE +/- 1.76, N = 3 298.14 MIN: 292.22 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch 1D - Data Type: f32 - Engine: CPU Intel I218-LM - Intel Core i7-5600U 5 10 15 20 25 SE +/- 0.28, N = 3 21.17 MIN: 19.63 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
Rodinia Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes select OpenCL, NVIDIA CUDA and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP Streamcluster Intel I218-LM - Intel Core i7-5600U 9 18 27 36 45 SE +/- 0.25, N = 3 40.62 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP CFD Solver Intel I218-LM - Intel Core i7-5600U 20 40 60 80 100 SE +/- 0.23, N = 3 101.30 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP Leukocyte Intel I218-LM - Intel Core i7-5600U 140 280 420 560 700 SE +/- 0.28, N = 3 625.91 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP HotSpot3D Intel I218-LM - Intel Core i7-5600U 50 100 150 200 250 SE +/- 0.92, N = 3 207.52 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenCL Myocyte Intel I218-LM - Intel Core i7-5600U 20 40 60 80 100 SE +/- 1.18, N = 3 110.08 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP LavaMD Intel I218-LM - Intel Core i7-5600U 400 800 1200 1600 2000 SE +/- 6.95, N = 3 1768.17 1. (CXX) g++ options: -O2 -lOpenCL
WireGuard + Linux Networking Stack Stress Test This is a benchmark of the WireGuard secure VPN tunnel and Linux networking stack stress test. The test runs on the local host but does require root permissions to run. The way it works is it creates three namespaces. ns0 has a loopback device. ns1 and ns2 each have wireguard devices. Those two wireguard devices send traffic through the loopback device of ns0. The end result of this is that tests wind up testing encryption and decryption at the same time -- a pretty CPU and scheduler-heavy workflow. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better WireGuard + Linux Networking Stack Stress Test Intel I218-LM - Intel Core i7-5600U 80 160 240 320 400 SE +/- 3.12, N = 3 385.05
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch All - Data Type: u8s8f32 - Engine: CPU Intel I218-LM - Intel Core i7-5600U 40 80 120 160 200 SE +/- 4.83, N = 15 195.81 MIN: 187.06 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
Intel I218-LM - Intel Core i7-5600U Processor: Intel Core i7-5600U @ 3.20GHz (2 Cores / 4 Threads), Motherboard: LENOVO 20BSCTO1WW (N14ET49W 1.27 BIOS), Chipset: Intel Broadwell-U-OPI, Memory: 8GB, Disk: 128GB SAMSUNG MZNTE128, Graphics: Intel HD 5500 3GB (950MHz), Audio: Intel Broadwell-U Audio, Network: Intel I218-LM + Intel 7265
OS: Ubuntu 20.04, Kernel: 5.4.0-40-generic (x86_64), Desktop: GNOME Shell 3.36.1, Display Server: X Server 1.20.8, Display Driver: modesetting 1.20.8, OpenGL: 4.6 Mesa 20.0.4, OpenCL: OpenCL 2.1, Compiler: GCC 9.3.0, File-System: ext4, Screen Resolution: 1920x1080
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersave - CPU Microcode: 0x2fPython Notes: Python 3.8.2Security Notes: itlb_multihit: KVM: Mitigation of Split huge pages + l1tf: Mitigation of PTE Inversion; VMX: conditional cache flushes SMT vulnerable + mds: Mitigation of Clear buffers; SMT vulnerable + meltdown: Mitigation of PTI + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + srbds: Mitigation of Microcode + tsx_async_abort: Mitigation of Clear buffers; SMT vulnerable
Testing initiated at 14 July 2020 13:25 by user phoronix.