rtx-2070
AMD Ryzen Threadripper 2990WX 32-Core testing with a ASUS ROG ZENITH EXTREME (1601 BIOS) and ASUS NVIDIA GeForce RTX 2070 8GB on Ubuntu 18.04 via the Phoronix Test Suite.
GeForce RTX 2070
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: acpi-cpufreq ondemand
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full AMD retpoline IBPB + SSB disabled via prctl and seccomp
ASUS NVIDIA GeForce RTX 2070
Processor: AMD Ryzen Threadripper 2990WX 32-Core @ 3.00GHz (32 Cores / 64 Threads), Motherboard: ASUS ROG ZENITH EXTREME (1601 BIOS), Chipset: AMD Family 17h, Memory: 32768MB, Disk: Samsung SSD 970 EVO 500GB, Graphics: ASUS NVIDIA GeForce RTX 2070 8GB (1410/7000MHz), Audio: Realtek ALC1220, Monitor: ASUS VP28U, Network: Intel I211 + Qualcomm Atheros QCA6174 802.11ac + Wilocity Wil6200 802.11ad
OS: Ubuntu 18.04, Kernel: 4.15.0-44-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 418.30, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.1.92, Vulkan: 1.1.95, Compiler: GCC 7.3.0, File-System: ext4, Screen Resolution: 3840x2160
Processor Notes: Scaling Governor: acpi-cpufreq ondemand
Security Notes: __user pointer sanitization + Full AMD retpoline IBPB + SSB disabled via prctl and seccomp
GLmark2
This is a test of any system-installed GLMark2 OpenGL benchmark. Learn more via the OpenBenchmarking.org test page.
ParaView
This test runs ParaView benchmarks: an open-source data analytics and visualization application. Learn more via the OpenBenchmarking.org test page.
GLmark2
This is a test of Linaro's glmark2 port, currently using the X11 OpenGL 2.0 target. GLmark2 is a basic OpenGL benchmark. Learn more via the OpenBenchmarking.org test page.
Xonotic
This is a benchmark of Xonotic, which is a fork of the DarkPlaces-based Nexuiz game. Development began in March of 2010 on the Xonotic game. Learn more via the OpenBenchmarking.org test page.
Unigine Valley
This test calculates the average frame-rate within the Valley demo for the Unigine engine, released in February 2013. This engine is extremely demanding on the system's graphics card. Unigine Valley relies upon an OpenGL 3 core profile context. Learn more via the OpenBenchmarking.org test page.
Unigine Superposition
This test calculates the average frame-rate within the Superposition demo for the Unigine engine, released in 2017. This engine is extremely demanding on the system's graphics card. Learn more via the OpenBenchmarking.org test page.
Unigine Heaven
This test calculates the average frame-rate within the Heaven demo for the Unigine engine. This engine is extremely demanding on the system's graphics card. Learn more via the OpenBenchmarking.org test page.
Tesseract
Tesseract is a fork of Cube 2 Sauerbraten with numerous graphics and game-play improvements. Tesseract has been in development since 2012 while its first release happened in May of 2014. Learn more via the OpenBenchmarking.org test page.
ParaView
This test runs ParaView benchmarks: an open-source data analytics and visualization application. Learn more via the OpenBenchmarking.org test page.
GeForce RTX 2070
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++ --enable-libmpx --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib --with-tune=generic --without-cuda-driver -v
Processor Notes: Scaling Governor: acpi-cpufreq ondemand
Python Notes: Python 2.7.15rc1 + Python 3.6.7
Security Notes: __user pointer sanitization + Full AMD retpoline IBPB + SSB disabled via prctl and seccomp
Testing initiated at 31 January 2019 10:18 by user pts.
ASUS NVIDIA GeForce RTX 2070
Processor: AMD Ryzen Threadripper 2990WX 32-Core @ 3.00GHz (32 Cores / 64 Threads), Motherboard: ASUS ROG ZENITH EXTREME (1601 BIOS), Chipset: AMD Family 17h, Memory: 32768MB, Disk: Samsung SSD 970 EVO 500GB, Graphics: ASUS NVIDIA GeForce RTX 2070 8GB (1410/7000MHz), Audio: Realtek ALC1220, Monitor: ASUS VP28U, Network: Intel I211 + Qualcomm Atheros QCA6174 802.11ac + Wilocity Wil6200 802.11ad
OS: Ubuntu 18.04, Kernel: 4.15.0-44-generic (x86_64), Desktop: GNOME Shell 3.28.3, Display Server: X Server 1.19.6, Display Driver: NVIDIA 418.30, OpenGL: 4.6.0, OpenCL: OpenCL 1.2 CUDA 10.1.92, Vulkan: 1.1.95, Compiler: GCC 7.3.0, File-System: ext4, Screen Resolution: 3840x2160
Processor Notes: Scaling Governor: acpi-cpufreq ondemand
Security Notes: __user pointer sanitization + Full AMD retpoline IBPB + SSB disabled via prctl and seccomp
Testing initiated at 31 January 2019 16:21 by user pts.