NVIDIA TITAN RTX Compute
NVIDIA TITAN RTX Linux GPGPU/compute benchmarks for a future article on Phoronix.com by Michael Larabel.
GTX 680
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: 2000GB SABRENT + Samsung SSD 970 EVO 250GB, Graphics: NVIDIA GeForce GTX 680 2GB (1006/3004MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel Connection
OS: Ubuntu 18.04, Kernel: 4.19.5-041905-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 415.23, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 7.3.0 + LLVM 6.0.0 + CUDA 10.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 1536
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp
GTX 780 Ti
Changed Graphics to NVIDIA GeForce GTX 780 Ti 3GB (875/3500MHz).
OpenCL Change: GPU Compute Cores: 2880
GTX 970
Changed Graphics to eVGA NVIDIA GeForce GTX 970 4GB (1163/3505MHz).
OpenCL Change: GPU Compute Cores: 1664
GTX 980
Changed Graphics to NVIDIA GeForce GTX 980 4GB (1126/3505MHz).
OpenCL Change: GPU Compute Cores: 2048
GTX 980 Ti
Changed Graphics to NVIDIA GeForce GTX 980 Ti 6GB (999/3505MHz).
OpenCL Change: GPU Compute Cores: 2816
GTX TITAN X GM200
Changed Graphics to NVIDIA GeForce GTX TITAN X 12GB (1001/3505MHz).
OpenCL Change: GPU Compute Cores: 3072
GTX 1060
Changed Graphics to NVIDIA GeForce GTX 1060 6GB (1506/4006MHz).
OpenCL Change: GPU Compute Cores: 1280
GTX 1070
Changed Graphics to NVIDIA GeForce GTX 1070 8GB (1506/4006MHz).
OpenCL Change: GPU Compute Cores: 1920
GTX 1080
Changed Graphics to NVIDIA GeForce GTX 1080 8GB (1607/5005MHz).
OpenCL Change: GPU Compute Cores: 2560
GTX 1080 Ti
Changed Graphics to NVIDIA GeForce GTX 1080 Ti 11GB (1480/5508MHz).
OpenCL Change: GPU Compute Cores: 3584
RTX 2080
Changed Graphics to Zotac NVIDIA GeForce RTX 2080 8GB (1515/7000MHz).
OpenCL Change: GPU Compute Cores: 2944
RTX 2080 Ti
Changed Graphics to NVIDIA GeForce RTX 2080 Ti 11GB (1350/7000MHz).
OpenCL Change: GPU Compute Cores: 4352
TITAN RTX
Changed Graphics to NVIDIA TITAN RTX 24GB (1350/7000MHz).
OpenCL Change: GPU Compute Cores: 4608
GTX-RTX-TITAN
Processor: 2 x Intel Xeon E5-4650L 0 @ 3.10GHz (16 Cores / 32 Threads), Motherboard: HP 158B (J63 v03.91 BIOS), Chipset: Intel Xeon E5/Core, Memory: 105472MB, Disk: 160GB Hitachi HTS72501, Graphics: AMD FirePro V5900 7GB, Audio: Realtek ALC262, Monitor: VGA, Network: Intel 82579LM + Intel 82574L
OS: Ubuntu 18.04, Kernel: 4.15.0-45-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6 + Wayland, Display Driver: modesetting 1.19.6, OpenGL: 4.3 Mesa 18.2.2 (LLVM 7.0.0), Compiler: GCC 7.3.0, File-System: ext4, Screen Resolution: 1920x1080
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate powersave
Graphics Notes: GLAMOR
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: KPTI + __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp + PTE Inversion
GROMACS
The CUDA version of the Gromacs molecular dynamics package. Learn more via the OpenBenchmarking.org test page.
OctaneBench
OctaneBench is a test of the OctaneRender on the GPU and requires the use of NVIDIA CUDA. Learn more via the OpenBenchmarking.org test page.
LuxMark
LuxMark is a multi-platform OpenGL benchmark using LuxRender. LuxMark supports targeting different OpenCL devices and has multiple scenes available for rendering. LuxMark is a fully open-source OpenCL program with real-world rendering examples. Learn more via the OpenBenchmarking.org test page.
FAHBench
FAHBench is a Folding@Home benchmark on the GPU. Learn more via the OpenBenchmarking.org test page.
NAMD CUDA
NAMD is a parallel molecular dynamics code designed for high-performance simulation of large biomolecular systems. NAMD was developed by the Theoretical and Computational Biophysics Group in the Beckman Institute for Advanced Science and Technology at the University of Illinois at Urbana-Champaign. This version of the NAMD test profile uses CUDA GPU acceleration. Learn more via the OpenBenchmarking.org test page.
Parboil
The Parboil Benchmarks from the IMPACT Research Group at University of Illinois are a set of throughput computing applications for looking at computing architecture and compilers. Parboil test-cases support OpenMP, OpenCL, and CUDA multi-processing environments. However, at this time the test profile is just making use of the OpenMP and OpenCL test workloads. Learn more via the OpenBenchmarking.org test page.
Rodinia
Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes the OpenCL and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.
Darktable
Darktable is an open-source photography / workflow application this will use any system-installed Darktable program or on Windows will automatically download the pre-built binary from the project. Learn more via the OpenBenchmarking.org test page.
clpeak
Clpeak is designed to test the peak capabilities of OpenCL devices. Learn more via the OpenBenchmarking.org test page.
Parboil
The Parboil Benchmarks from the IMPACT Research Group at University of Illinois are a set of throughput computing applications for looking at computing architecture and compilers. Parboil test-cases support OpenMP, OpenCL, and CUDA multi-processing environments. However, at this time the test profile is just making use of the OpenMP and OpenCL test workloads. Learn more via the OpenBenchmarking.org test page.
Mixbench
A benchmark suite for GPUs on mixed operational intensity kernels. Learn more via the OpenBenchmarking.org test page.
System Power Consumption Monitor
GPU Temperature Monitor
GROMACS
LuxMark
FAHBench
NAMD CUDA
OctaneBench
GTX 680
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: 2000GB SABRENT + Samsung SSD 970 EVO 250GB, Graphics: NVIDIA GeForce GTX 680 2GB (1006/3004MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel Connection
OS: Ubuntu 18.04, Kernel: 4.19.5-041905-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 415.23, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 7.3.0 + LLVM 6.0.0 + CUDA 10.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 1536
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp
Testing initiated at 23 December 2018 13:11 by user pts.
GTX 780 Ti
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: 2000GB SABRENT + Samsung SSD 970 EVO 250GB, Graphics: NVIDIA GeForce GTX 780 Ti 3GB (875/3500MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel Connection
OS: Ubuntu 18.04, Kernel: 4.19.5-041905-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 415.23, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 7.3.0 + LLVM 6.0.0 + CUDA 10.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 2880
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp
Testing initiated at 22 December 2018 21:54 by user pts.
GTX 970
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: 2000GB SABRENT + Samsung SSD 970 EVO 250GB, Graphics: eVGA NVIDIA GeForce GTX 970 4GB (1163/3505MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel Connection
OS: Ubuntu 18.04, Kernel: 4.19.5-041905-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 415.23, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 7.3.0 + LLVM 6.0.0 + CUDA 10.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 1664
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp
Testing initiated at 23 December 2018 19:20 by user pts.
GTX 980
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: 2000GB SABRENT + Samsung SSD 970 EVO 250GB, Graphics: NVIDIA GeForce GTX 980 4GB (1126/3505MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel Connection
OS: Ubuntu 18.04, Kernel: 4.19.5-041905-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 415.23, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 7.3.0 + LLVM 6.0.0 + CUDA 10.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 2048
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp
Testing initiated at 22 December 2018 19:20 by user pts.
GTX 980 Ti
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: 2000GB SABRENT + Samsung SSD 970 EVO 250GB, Graphics: NVIDIA GeForce GTX 980 Ti 6GB (999/3505MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel Connection
OS: Ubuntu 18.04, Kernel: 4.19.5-041905-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 415.23, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 7.3.0 + LLVM 6.0.0 + CUDA 10.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 2816
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp
Testing initiated at 23 December 2018 14:48 by user pts.
GTX TITAN X GM200
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: 2000GB SABRENT + Samsung SSD 970 EVO 250GB, Graphics: NVIDIA GeForce GTX TITAN X 12GB (1001/3505MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel Connection
OS: Ubuntu 18.04, Kernel: 4.19.5-041905-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 415.23, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 7.3.0 + LLVM 6.0.0 + CUDA 10.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 3072
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp
Testing initiated at 22 December 2018 18:20 by user pts.
GTX 1060
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: 2000GB SABRENT + Samsung SSD 970 EVO 250GB, Graphics: NVIDIA GeForce GTX 1060 6GB (1506/4006MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel Connection
OS: Ubuntu 18.04, Kernel: 4.19.5-041905-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 415.23, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 7.3.0 + LLVM 6.0.0 + CUDA 10.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 1280
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp
Testing initiated at 23 December 2018 07:14 by user pts.
GTX 1070
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: 2000GB SABRENT + Samsung SSD 970 EVO 250GB, Graphics: NVIDIA GeForce GTX 1070 8GB (1506/4006MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel Connection
OS: Ubuntu 18.04, Kernel: 4.19.5-041905-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 415.23, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 7.3.0 + LLVM 6.0.0 + CUDA 10.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 1920
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp
Testing initiated at 23 December 2018 12:10 by user pts.
GTX 1080
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: 2000GB SABRENT + Samsung SSD 970 EVO 250GB, Graphics: NVIDIA GeForce GTX 1080 8GB (1607/5005MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel Connection
OS: Ubuntu 18.04, Kernel: 4.19.5-041905-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 415.23, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 7.3.0 + LLVM 6.0.0 + CUDA 10.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 2560
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp
Testing initiated at 22 December 2018 20:18 by user pts.
GTX 1080 Ti
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: 2000GB SABRENT + Samsung SSD 970 EVO 250GB, Graphics: NVIDIA GeForce GTX 1080 Ti 11GB (1480/5508MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel Connection
OS: Ubuntu 18.04, Kernel: 4.19.5-041905-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 415.23, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 7.3.0 + LLVM 6.0.0 + CUDA 10.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 3584
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp
Testing initiated at 23 December 2018 10:49 by user pts.
RTX 2080
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: 2000GB SABRENT + Samsung SSD 970 EVO 250GB, Graphics: Zotac NVIDIA GeForce RTX 2080 8GB (1515/7000MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel Connection
OS: Ubuntu 18.04, Kernel: 4.19.5-041905-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 415.23, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 7.3.0 + LLVM 6.0.0 + CUDA 10.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 2944
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp
Testing initiated at 23 December 2018 09:09 by user pts.
RTX 2080 Ti
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: 2000GB SABRENT + Samsung SSD 970 EVO 250GB, Graphics: NVIDIA GeForce RTX 2080 Ti 11GB (1350/7000MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel Connection
OS: Ubuntu 18.04, Kernel: 4.19.5-041905-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 415.23, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 7.3.0 + LLVM 6.0.0 + CUDA 10.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 4352
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp
Testing initiated at 23 December 2018 16:28 by user pts.
TITAN RTX
Processor: Intel Core i9-9900K @ 5.00GHz (8 Cores / 16 Threads), Motherboard: ASUS PRIME Z390-A (0602 BIOS), Chipset: Intel Cannon Lake PCH Shared SRAM, Memory: 16384MB, Disk: 2000GB SABRENT + Samsung SSD 970 EVO 250GB, Graphics: NVIDIA TITAN RTX 24GB (1350/7000MHz), Audio: Realtek ALC1220, Monitor: Acer B286HK, Network: Intel Connection
OS: Ubuntu 18.04, Kernel: 4.19.5-041905-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 415.23, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.0.132, Vulkan: 1.1.84, Compiler: GCC 7.3.0 + LLVM 6.0.0 + CUDA 10.0, File-System: ext4, Screen Resolution: 3840x2160
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate performance
OpenCL Notes: GPU Compute Cores: 4608
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp
Testing initiated at 22 December 2018 09:33 by user pts.
GTX-RTX-TITAN
Processor: 2 x Intel Xeon E5-4650L 0 @ 3.10GHz (16 Cores / 32 Threads), Motherboard: HP 158B (J63 v03.91 BIOS), Chipset: Intel Xeon E5/Core, Memory: 105472MB, Disk: 160GB Hitachi HTS72501, Graphics: AMD FirePro V5900 7GB, Audio: Realtek ALC262, Monitor: VGA, Network: Intel 82579LM + Intel 82574L
OS: Ubuntu 18.04, Kernel: 4.15.0-45-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6 + Wayland, Display Driver: modesetting 1.19.6, OpenGL: 4.3 Mesa 18.2.2 (LLVM 7.0.0), Compiler: GCC 7.3.0, File-System: ext4, Screen Resolution: 1920x1080
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: intel_pstate powersave
Graphics Notes: GLAMOR
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: KPTI + __user pointer sanitization + Full generic retpoline IBPB IBRS_FW + SSB disabled via prctl and seccomp + PTE Inversion
Testing initiated at 28 February 2019 11:35 by user fwin.