Intel Core i9-10900X testing with a Gigabyte X299-WU8-CF (F5h BIOS) and Gigabyte NVIDIA GeForce RTX 3080 10GB on Ubuntu 22.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2308215-NE-SFSDF602926
Jump To Table - Results


Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs


Show Detailed System Result Table

Run Management

View Logs
Performance Per
Intel Core i9-10900X - Gigabyte NVIDIA GeForce RTX
August 21
  15 Minutes

sfsdfOpenBenchmarking.orgPhoronix Test SuiteIntel Core i9-10900X @ 8.00GHz (10 Cores / 20 Threads)Gigabyte X299-WU8-CF (F5h BIOS)Intel Sky Lake-E DMI3 Registers128GB500GB SHGP31-500GMGigabyte NVIDIA GeForce RTX 3080 10GBRealtek ALC1220EV3237Intel I219-LM + Intel I210Ubuntu (x86_64)GNOME Shell 42.9X Server 525. 3.0 CUDA 12.0.1511.3.224GCC 11.4.0ext43840x2160ProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerDisplay DriverOpenGLOpenCLVulkanCompilerFile-SystemScreen ResolutionSfsdf BenchmarksSystem Logs- Transparent Huge Pages: madvise- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-XeT9lY/gcc-11-11.4.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-XeT9lY/gcc-11-11.4.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x5003604 - GPU Compute Cores: 8704- itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Mitigation of Clear buffers; SMT vulnerable + retbleed: Mitigation of Enhanced IBRS + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Not affected + tsx_async_abort: Mitigation of TSX disabled


LuxMark is a multi-platform OpenGL benchmark using LuxRender. LuxMark supports targeting different OpenCL devices and has multiple scenes available for rendering. LuxMark is a fully open-source OpenCL program with real-world rendering examples. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgScore, More Is BetterLuxMark 3.1OpenCL Device: GPU - Scene: MicrophoneIntel Core i9-10900X - Gigabyte NVIDIA GeForce RTX30K60K90K120K150KSE +/- 1268.36, N = 7138193