Xeon Gold Benchmark Fun 2 x Intel Xeon Gold 6138 testing with a TYAN S7106 (V2.00.B20 BIOS) and llvmpipe 93GB on Ubuntu 18.04 via the Phoronix Test Suite.
Compare your own system(s) to this result file with the
Phoronix Test Suite by running the command:
phoronix-test-suite benchmark 1910060-AS-XEONGOLDB99 2 x Intel Xeon Gold 6138 Processor: 2 x Intel Xeon Gold 6138 @ 3.70GHz (40 Cores / 80 Threads), Motherboard: TYAN S7106 (V2.00.B20 BIOS), Chipset: Intel Sky Lake-E DMI3 Registers, Memory: 96256MB, Disk: 500GB Samsung SSD 860, Graphics: llvmpipe 93GB, Monitor: VE228, Network: 2 x Intel I210 + 2 x QLogic cLOM8214 1/10GbE
OS: Ubuntu 18.04, Kernel: 5.0.0-29-generic (x86_64), Desktop: GNOME Shell 3.28.4, Display Server: X Server 1.20.4, Display Driver: modesetting 1.20.4, OpenGL: 3.3 Mesa 19.0.8 (LLVM 8.0 256 bits), Compiler: GCC 7.4.0, File-System: ext4, Screen Resolution: 1920x1080
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersaveSecurity Notes: l1tf: Mitigation of PTE Inversion; VMX: conditional cache flushes SMT vulnerable + mds: Mitigation of Clear buffers; SMT vulnerable + meltdown: Mitigation of PTI + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling
Xeon Gold Benchmark Fun OpenBenchmarking.org Phoronix Test Suite 2 x Intel Xeon Gold 6138 @ 3.70GHz (40 Cores / 80 Threads) TYAN S7106 (V2.00.B20 BIOS) Intel Sky Lake-E DMI3 Registers 96256MB 500GB Samsung SSD 860 llvmpipe 93GB VE228 2 x Intel I210 + 2 x QLogic cLOM8214 1/10GbE Ubuntu 18.04 5.0.0-29-generic (x86_64) GNOME Shell 3.28.4 X Server 1.20.4 modesetting 1.20.4 3.3 Mesa 19.0.8 (LLVM 8.0 256 bits) GCC 7.4.0 ext4 1920x1080 Processor Motherboard Chipset Memory Disk Graphics Monitor Network OS Kernel Desktop Display Server Display Driver OpenGL Compiler File-System Screen Resolution Xeon Gold Benchmark Fun Performance System Logs - --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v - Scaling Governor: intel_pstate powersave - l1tf: Mitigation of PTE Inversion; VMX: conditional cache flushes SMT vulnerable + mds: Mitigation of Clear buffers; SMT vulnerable + meltdown: Mitigation of PTI + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling
Xeon Gold Benchmark Fun mkl-dnn: Deconvolution Batch deconv_all - f32 lczero: BLAS lczero: Rand mkl-dnn: Deconvolution Batch deconv_1d - u8s8f32 libgav1: Chimera 1080p mkl-dnn: Convolution Batch conv_googlenet_v3 - bf16bf16bf16 libgav1: Summer Nature 1080p mkl-dnn: Deconvolution Batch deconv_3d - bf16bf16bf16 mkl-dnn: Convolution Batch conv_googlenet_v3 - u8s8f32 mkl-dnn: Convolution Batch conv_alexnet - u8s8f32 mkl-dnn: Convolution Batch conv_googlenet_v3 - f32 mkl-dnn: Convolution Batch conv_3d - f32 mkl-dnn: IP Batch 1D - f32 mkl-dnn: IP Batch 1D - bf16bf16bf16 mkl-dnn: IP Batch All - bf16bf16bf16 mkl-dnn: IP Batch All - u8s8f32 mkl-dnn: IP Batch 1D - u8s8f32 mkl-dnn: IP Batch All - f32 libgav1: Summer Nature 4K mkl-dnn: Convolution Batch conv_all - f32 mkl-dnn: Deconvolution Batch deconv_3d - u8s8f32 mkl-dnn: Convolution Batch conv_3d - u8s8f32 mkl-dnn: Convolution Batch conv_3d - bf16bf16bf16 mkl-dnn: Deconvolution Batch deconv_1d - f32 mkl-dnn: Convolution Batch conv_all - bf16bf16bf16 mkl-dnn: Deconvolution Batch deconv_3d - f32 mkl-dnn: Deconvolution Batch deconv_1d - bf16bf16bf16 mkl-dnn: Convolution Batch conv_alexnet - f32 mkl-dnn: Convolution Batch conv_alexnet - bf16bf16bf16 mkl-dnn: Convolution Batch conv_all - u8s8f32 mkl-dnn: Deconvolution Batch deconv_all - bf16bf16bf16 mkl-dnn: Recurrent Neural Network Training - f32 libgav1: Chimera 1080p 10-bit pgbench: Buffer Test - Normal Load - Read Only pgbench: Buffer Test - Normal Load - Read Write pgbench: Buffer Test - Single Thread - Read Only pgbench: Buffer Test - Single Thread - Read Write pgbench: Buffer Test - Heavy Contention - Read Only pgbench: Buffer Test - Heavy Contention - Read Write 2 x Intel Xeon Gold 6138 1120.20 9.03 40306.50 0.89 24.78 152.96 35.87 7.11 25.39 60.03 39.81 6.15 2.49 4.78 154.75 4.58 2.14 11.74 12.01 851.13 3950.66 4556.54 12.80 1.74 3108.26 1.94 6.39 86.68 567.61 2515.59 2756.70 246.86 11.76 573225.58 4739.81 18109.92 256.04 608561.84 4592.85 OpenBenchmarking.org
MKL-DNN DNNL This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Deconvolution Batch deconv_all - Data Type: f32 2 x Intel Xeon Gold 6138 200 400 600 800 1000 SE +/- 0.70, N = 3 1120.20 MIN: 1109.47 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
MKL-DNN DNNL This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32 2 x Intel Xeon Gold 6138 0.2003 0.4006 0.6009 0.8012 1.0015 SE +/- 0.01, N = 3 0.89 MIN: 0.85 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
MKL-DNN DNNL This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Convolution Batch conv_googlenet_v3 - Data Type: bf16bf16bf16 2 x Intel Xeon Gold 6138 30 60 90 120 150 SE +/- 0.69, N = 3 152.96 MIN: 150.11 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
MKL-DNN DNNL This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Deconvolution Batch deconv_3d - Data Type: bf16bf16bf16 2 x Intel Xeon Gold 6138 2 4 6 8 10 SE +/- 0.01, N = 3 7.11 MIN: 7.04 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Convolution Batch conv_googlenet_v3 - Data Type: u8s8f32 2 x Intel Xeon Gold 6138 6 12 18 24 30 SE +/- 0.10, N = 3 25.39 MIN: 24.46 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Convolution Batch conv_alexnet - Data Type: u8s8f32 2 x Intel Xeon Gold 6138 13 26 39 52 65 SE +/- 0.09, N = 3 60.03 MIN: 57.37 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Convolution Batch conv_googlenet_v3 - Data Type: f32 2 x Intel Xeon Gold 6138 9 18 27 36 45 SE +/- 0.33, N = 3 39.81 MIN: 37.31 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Convolution Batch conv_3d - Data Type: f32 2 x Intel Xeon Gold 6138 2 4 6 8 10 SE +/- 0.01, N = 3 6.15 MIN: 5.98 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: IP Batch 1D - Data Type: f32 2 x Intel Xeon Gold 6138 0.5603 1.1206 1.6809 2.2412 2.8015 SE +/- 0.01, N = 3 2.49 MIN: 2.19 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: IP Batch 1D - Data Type: bf16bf16bf16 2 x Intel Xeon Gold 6138 1.0755 2.151 3.2265 4.302 5.3775 SE +/- 0.00, N = 3 4.78 MIN: 4.46 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: IP Batch All - Data Type: bf16bf16bf16 2 x Intel Xeon Gold 6138 30 60 90 120 150 SE +/- 0.41, N = 3 154.75 MIN: 116.15 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: IP Batch All - Data Type: u8s8f32 2 x Intel Xeon Gold 6138 1.0305 2.061 3.0915 4.122 5.1525 SE +/- 0.02, N = 3 4.58 MIN: 4.28 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: IP Batch 1D - Data Type: u8s8f32 2 x Intel Xeon Gold 6138 0.4815 0.963 1.4445 1.926 2.4075 SE +/- 0.01, N = 3 2.14 MIN: 1.98 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: IP Batch All - Data Type: f32 2 x Intel Xeon Gold 6138 3 6 9 12 15 SE +/- 0.03, N = 3 11.74 MIN: 11.33 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
MKL-DNN DNNL This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Convolution Batch conv_all - Data Type: f32 2 x Intel Xeon Gold 6138 200 400 600 800 1000 SE +/- 0.32, N = 3 851.13 MIN: 837.01 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32 2 x Intel Xeon Gold 6138 800 1600 2400 3200 4000 SE +/- 5.13, N = 3 3950.66 MIN: 3935.28 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Convolution Batch conv_3d - Data Type: u8s8f32 2 x Intel Xeon Gold 6138 1000 2000 3000 4000 5000 SE +/- 3.70, N = 3 4556.54 MIN: 4539.32 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Convolution Batch conv_3d - Data Type: bf16bf16bf16 2 x Intel Xeon Gold 6138 3 6 9 12 15 SE +/- 0.01, N = 3 12.80 MIN: 12.61 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Deconvolution Batch deconv_1d - Data Type: f32 2 x Intel Xeon Gold 6138 0.3915 0.783 1.1745 1.566 1.9575 SE +/- 0.00, N = 3 1.74 MIN: 1.67 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Convolution Batch conv_all - Data Type: bf16bf16bf16 2 x Intel Xeon Gold 6138 700 1400 2100 2800 3500 SE +/- 0.74, N = 3 3108.26 MIN: 3096.57 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Deconvolution Batch deconv_3d - Data Type: f32 2 x Intel Xeon Gold 6138 0.4365 0.873 1.3095 1.746 2.1825 SE +/- 0.00, N = 3 1.94 MIN: 1.89 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Deconvolution Batch deconv_1d - Data Type: bf16bf16bf16 2 x Intel Xeon Gold 6138 2 4 6 8 10 SE +/- 0.00, N = 3 6.39 MIN: 6.29 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Convolution Batch conv_alexnet - Data Type: f32 2 x Intel Xeon Gold 6138 20 40 60 80 100 SE +/- 0.15, N = 3 86.68 MIN: 83.65 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Convolution Batch conv_alexnet - Data Type: bf16bf16bf16 2 x Intel Xeon Gold 6138 120 240 360 480 600 SE +/- 1.75, N = 3 567.61 MIN: 562.42 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Convolution Batch conv_all - Data Type: u8s8f32 2 x Intel Xeon Gold 6138 500 1000 1500 2000 2500 SE +/- 10.73, N = 3 2515.59 MIN: 2490.58 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Deconvolution Batch deconv_all - Data Type: bf16bf16bf16 2 x Intel Xeon Gold 6138 600 1200 1800 2400 3000 SE +/- 1.64, N = 3 2756.70 MIN: 2723.66 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better MKL-DNN DNNL 1.1 Harness: Recurrent Neural Network Training - Data Type: f32 2 x Intel Xeon Gold 6138 50 100 150 200 250 SE +/- 3.23, N = 4 246.86 MIN: 232.6 1. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl
OpenBenchmarking.org TPS, More Is Better PostgreSQL pgbench 12.0 Scaling: Buffer Test - Test: Normal Load - Mode: Read Write 2 x Intel Xeon Gold 6138 1000 2000 3000 4000 5000 SE +/- 45.05, N = 12 4739.81 1. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -lcrypt -ldl -lm
OpenBenchmarking.org TPS, More Is Better PostgreSQL pgbench 12.0 Scaling: Buffer Test - Test: Single Thread - Mode: Read Only 2 x Intel Xeon Gold 6138 4K 8K 12K 16K 20K SE +/- 60.30, N = 3 18109.92 1. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -lcrypt -ldl -lm
OpenBenchmarking.org TPS, More Is Better PostgreSQL pgbench 12.0 Scaling: Buffer Test - Test: Single Thread - Mode: Read Write 2 x Intel Xeon Gold 6138 60 120 180 240 300 SE +/- 0.48, N = 3 256.04 1. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -lcrypt -ldl -lm
OpenBenchmarking.org TPS, More Is Better PostgreSQL pgbench 12.0 Scaling: Buffer Test - Test: Heavy Contention - Mode: Read Only 2 x Intel Xeon Gold 6138 130K 260K 390K 520K 650K SE +/- 3362.73, N = 3 608561.84 1. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -lcrypt -ldl -lm
OpenBenchmarking.org TPS, More Is Better PostgreSQL pgbench 12.0 Scaling: Buffer Test - Test: Heavy Contention - Mode: Read Write 2 x Intel Xeon Gold 6138 1000 2000 3000 4000 5000 SE +/- 76.41, N = 3 4592.85 1. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -lcrypt -ldl -lm
2 x Intel Xeon Gold 6138 Processor: 2 x Intel Xeon Gold 6138 @ 3.70GHz (40 Cores / 80 Threads), Motherboard: TYAN S7106 (V2.00.B20 BIOS), Chipset: Intel Sky Lake-E DMI3 Registers, Memory: 96256MB, Disk: 500GB Samsung SSD 860, Graphics: llvmpipe 93GB, Monitor: VE228, Network: 2 x Intel I210 + 2 x QLogic cLOM8214 1/10GbE
OS: Ubuntu 18.04, Kernel: 5.0.0-29-generic (x86_64), Desktop: GNOME Shell 3.28.4, Display Server: X Server 1.20.4, Display Driver: modesetting 1.20.4, OpenGL: 3.3 Mesa 19.0.8 (LLVM 8.0 256 bits), Compiler: GCC 7.4.0, File-System: ext4, Screen Resolution: 1920x1080
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -vProcessor Notes: Scaling Governor: intel_pstate powersaveSecurity Notes: l1tf: Mitigation of PTE Inversion; VMX: conditional cache flushes SMT vulnerable + mds: Mitigation of Clear buffers; SMT vulnerable + meltdown: Mitigation of PTI + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling
Testing initiated at 6 October 2019 09:37 by user phoronix.