NVIDIA Linux GPU Compute
NVIDIA Linux GPU computing benchmarks by Michael Larabel.
GTX 1060
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: Samsung SSD 970 EVO 250GB + 2000GB SABRENT, Graphics: NVIDIA GeForce GTX 1060 6GB (1506/4006MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel I219-V
OS: Ubuntu 18.10, Kernel: 4.20.3-042003-generic (x86_64), Desktop: GNOME Shell 3.30.1, Display Server: X Server 1.20.1, Display Driver: NVIDIA 415.27, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 8.2.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 1280
Python Notes: Python 2.7.15+ + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + SSB disabled via prctl and seccomp
GTX 1070
Changed Graphics to NVIDIA GeForce GTX 1070 8GB (1506/4006MHz).
OpenCL Change: GPU Compute Cores: 1920
GTX 1070 Ti
Changed Graphics to Zotac NVIDIA GeForce GTX 1070 Ti 8GB (1607/4006MHz).
OpenCL Change: GPU Compute Cores: 2432
GTX 1080
Changed Graphics to NVIDIA GeForce GTX 1080 8GB (1607/5005MHz).
OpenCL Change: GPU Compute Cores: 2560
GTX 1080 Ti
Changed Graphics to NVIDIA GeForce GTX 1080 Ti 11GB (1480/5508MHz).
OpenCL Change: GPU Compute Cores: 3584
RTX 2060
Changed Graphics to NVIDIA GeForce RTX 2060 6GB (1365/7000MHz).
OpenCL Change: GPU Compute Cores: 1920
RTX 2070
Changed Graphics to ASUS NVIDIA GeForce RTX 2070 8GB (1410/7000MHz).
OpenCL Change: GPU Compute Cores: 2304
RTX 2080
Changed Graphics to Zotac NVIDIA GeForce RTX 2080 8GB (1515/7000MHz).
OpenCL Change: GPU Compute Cores: 2944
RTX 2080 Ti
Changed Graphics to NVIDIA GeForce RTX 2080 Ti 11GB (1350/7000MHz).
OpenCL Change: GPU Compute Cores: 4352
TITAN RTX
Changed Graphics to NVIDIA TITAN RTX 24GB (1350/7000MHz).
OpenCL Change: GPU Compute Cores: 4608
p100-host
Processor: 2 x Intel Xeon E5-2680 v3 @ 2.50GHz (24 Cores / 48 Threads), Motherboard: HP ProLiant DL380 Gen9 (P89 BIOS), Chipset: Intel Xeon E7 v3/Xeon, Memory: 16 x 16384 MB DDR4-2133MT/s 752369-081, Disk: 600GB LOGICAL VOLUME + 3000GB LOGICAL VOLUME, Graphics: Matrox MGA G200EH, Network: Broadcom Limited NetXtreme BCM5719 Gigabit PCIe
OS: Ubuntu 18.04, Kernel: 4.15.0-45-generic (x86_64), Compiler: GCC 7.3.0 + CUDA 9.1, File-System: ext4, Screen Resolution: 1024x768
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: pcc-cpufreq ondemand
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: KPTI + __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp + PTE Inversion; VMX: conditional cache flushes SMT vulnerable
LeelaChessZero
LeelaChessZero (lc0 / lczero) is a chess engine automated vian neural networks. This test profile can be used for OpenCL, CUDA + cuDNN, and BLAS (CPU-based) benchmarking. Learn more via the OpenBenchmarking.org test page.
PlaidML
This test profile uses PlaidML deep learning framework for offering up various benchmarks. Learn more via the OpenBenchmarking.org test page.
NAMD CUDA
NAMD is a parallel molecular dynamics code designed for high-performance simulation of large biomolecular systems. NAMD was developed by the Theoretical and Computational Biophysics Group in the Beckman Institute for Advanced Science and Technology at the University of Illinois at Urbana-Champaign. This version of the NAMD test profile uses CUDA GPU acceleration. Learn more via the OpenBenchmarking.org test page.
Darktable
Darktable is an open-source photography / workflow application this will use any system-installed Darktable program or on Windows will automatically download the pre-built binary from the project. Learn more via the OpenBenchmarking.org test page.
LuxMark
LuxMark is a multi-platform OpenGL benchmark using LuxRender. LuxMark supports targeting different OpenCL devices and has multiple scenes available for rendering. LuxMark is a fully open-source OpenCL program with real-world rendering examples. Learn more via the OpenBenchmarking.org test page.
cl-mem
A basic OpenCL memory benchmark. Learn more via the OpenBenchmarking.org test page.
clpeak
SHOC Scalable HeterOgeneous Computing
The CUDA and OpenCL version of Vetter's Scalable HeterOgeneous Computing benchmark suite. Learn more via the OpenBenchmarking.org test page.
JuliaGPU
JuliaGPU is an OpenCL benchmark with this version containing various PTS-specific enhancements. Learn more via the OpenBenchmarking.org test page.
Rodinia
Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes the OpenCL and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.
Meta Performance Per Watt
GPU Temperature Monitor
System Power Consumption Monitor
OctaneBench
OctaneBench is a test of the OctaneRender on the GPU and requires the use of NVIDIA CUDA. Learn more via the OpenBenchmarking.org test page.
Chaos Group V-RAY
This is a test of Chaos Group's V-RAY benchmark. Learn more via the OpenBenchmarking.org test page.
Darktable
Darktable is an open-source photography / workflow application this will use any system-installed Darktable program or on Windows will automatically download the pre-built binary from the project. Learn more via the OpenBenchmarking.org test page.
GTX 1060
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: Samsung SSD 970 EVO 250GB + 2000GB SABRENT, Graphics: NVIDIA GeForce GTX 1060 6GB (1506/4006MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel I219-V
OS: Ubuntu 18.10, Kernel: 4.20.3-042003-generic (x86_64), Desktop: GNOME Shell 3.30.1, Display Server: X Server 1.20.1, Display Driver: NVIDIA 415.27, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 8.2.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 1280
Python Notes: Python 2.7.15+ + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + SSB disabled via prctl and seccomp
Testing initiated at 27 January 2019 18:10 by user phoronix.
GTX 1070
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: Samsung SSD 970 EVO 250GB + 2000GB SABRENT, Graphics: NVIDIA GeForce GTX 1070 8GB (1506/4006MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel I219-V
OS: Ubuntu 18.10, Kernel: 4.20.3-042003-generic (x86_64), Desktop: GNOME Shell 3.30.1, Display Server: X Server 1.20.1, Display Driver: NVIDIA 415.27, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 8.2.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 1920
Python Notes: Python 2.7.15+ + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + SSB disabled via prctl and seccomp
Testing initiated at 25 January 2019 07:17 by user phoronix.
GTX 1070 Ti
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: Samsung SSD 970 EVO 250GB + 2000GB SABRENT, Graphics: Zotac NVIDIA GeForce GTX 1070 Ti 8GB (1607/4006MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel I219-V
OS: Ubuntu 18.10, Kernel: 4.20.3-042003-generic (x86_64), Desktop: GNOME Shell 3.30.1, Display Server: X Server 1.20.1, Display Driver: NVIDIA 415.27, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 8.2.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 2432
Python Notes: Python 2.7.15+ + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + SSB disabled via prctl and seccomp
Testing initiated at 25 January 2019 12:27 by user phoronix.
GTX 1080
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: Samsung SSD 970 EVO 250GB + 2000GB SABRENT, Graphics: NVIDIA GeForce GTX 1080 8GB (1607/5005MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel I219-V
OS: Ubuntu 18.10, Kernel: 4.20.3-042003-generic (x86_64), Desktop: GNOME Shell 3.30.1, Display Server: X Server 1.20.1, Display Driver: NVIDIA 415.27, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 8.2.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 2560
Python Notes: Python 2.7.15+ + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + SSB disabled via prctl and seccomp
Testing initiated at 25 January 2019 05:13 by user phoronix.
GTX 1080 Ti
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: Samsung SSD 970 EVO 250GB + 2000GB SABRENT, Graphics: NVIDIA GeForce GTX 1080 Ti 11GB (1480/5508MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel I219-V
OS: Ubuntu 18.10, Kernel: 4.20.3-042003-generic (x86_64), Desktop: GNOME Shell 3.30.1, Display Server: X Server 1.20.1, Display Driver: NVIDIA 415.27, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 8.2.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 3584
Python Notes: Python 2.7.15+ + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + SSB disabled via prctl and seccomp
Testing initiated at 25 January 2019 17:40 by user phoronix.
RTX 2060
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: Samsung SSD 970 EVO 250GB + 2000GB SABRENT, Graphics: NVIDIA GeForce RTX 2060 6GB (1365/7000MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel I219-V
OS: Ubuntu 18.10, Kernel: 4.20.3-042003-generic (x86_64), Desktop: GNOME Shell 3.30.1, Display Server: X Server 1.20.1, Display Driver: NVIDIA 415.27, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 8.2.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 1920
Python Notes: Python 2.7.15+ + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + SSB disabled via prctl and seccomp
Testing initiated at 27 January 2019 08:45 by user phoronix.
RTX 2070
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: Samsung SSD 970 EVO 250GB + 2000GB SABRENT, Graphics: ASUS NVIDIA GeForce RTX 2070 8GB (1410/7000MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel I219-V
OS: Ubuntu 18.10, Kernel: 4.20.3-042003-generic (x86_64), Desktop: GNOME Shell 3.30.1, Display Server: X Server 1.20.1, Display Driver: NVIDIA 415.27, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 8.2.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 2304
Python Notes: Python 2.7.15+ + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + SSB disabled via prctl and seccomp
Testing initiated at 27 January 2019 10:46 by user phoronix.
RTX 2080
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: Samsung SSD 970 EVO 250GB + 2000GB SABRENT, Graphics: Zotac NVIDIA GeForce RTX 2080 8GB (1515/7000MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel I219-V
OS: Ubuntu 18.10, Kernel: 4.20.3-042003-generic (x86_64), Desktop: GNOME Shell 3.30.1, Display Server: X Server 1.20.1, Display Driver: NVIDIA 415.27, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 8.2.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 2944
Python Notes: Python 2.7.15+ + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + SSB disabled via prctl and seccomp
Testing initiated at 26 January 2019 15:41 by user phoronix.
RTX 2080 Ti
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: Samsung SSD 970 EVO 250GB + 2000GB SABRENT, Graphics: NVIDIA GeForce RTX 2080 Ti 11GB (1350/7000MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel I219-V
OS: Ubuntu 18.10, Kernel: 4.20.3-042003-generic (x86_64), Desktop: GNOME Shell 3.30.1, Display Server: X Server 1.20.1, Display Driver: NVIDIA 415.27, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 8.2.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 4352
Python Notes: Python 2.7.15+ + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + SSB disabled via prctl and seccomp
Testing initiated at 26 January 2019 10:45 by user phoronix.
TITAN RTX
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: Samsung SSD 970 EVO 250GB + 2000GB SABRENT, Graphics: NVIDIA TITAN RTX 24GB (1350/7000MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel I219-V
OS: Ubuntu 18.10, Kernel: 4.20.3-042003-generic (x86_64), Desktop: GNOME Shell 3.30.1, Display Server: X Server 1.20.1, Display Driver: NVIDIA 415.27, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 8.2.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 4608
Python Notes: Python 2.7.15+ + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + SSB disabled via prctl and seccomp
Testing initiated at 26 January 2019 05:39 by user phoronix.
p100-host
Processor: 2 x Intel Xeon E5-2680 v3 @ 2.50GHz (24 Cores / 48 Threads), Motherboard: HP ProLiant DL380 Gen9 (P89 BIOS), Chipset: Intel Xeon E7 v3/Xeon, Memory: 16 x 16384 MB DDR4-2133MT/s 752369-081, Disk: 600GB LOGICAL VOLUME + 3000GB LOGICAL VOLUME, Graphics: Matrox MGA G200EH, Network: Broadcom Limited NetXtreme BCM5719 Gigabit PCIe
OS: Ubuntu 18.04, Kernel: 4.15.0-45-generic (x86_64), Compiler: GCC 7.3.0 + CUDA 9.1, File-System: ext4, Screen Resolution: 1024x768
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: pcc-cpufreq ondemand
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: KPTI + __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp + PTE Inversion; VMX: conditional cache flushes SMT vulnerable
Testing initiated at 5 February 2019 09:47 by user root.