ddds

Intel Core i7-1280P testing with a MSI MS-14C6 (E14C6IMS.115 BIOS) and MSI Intel ADL GT2 15GB on Ubuntu 23.10 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2310187-NE-DDDS2145567
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results
Show Result Confidence Charts

Limit displaying results to tests within:

Creator Workloads 4 Tests
Game Development 2 Tests
Multi-Core 4 Tests
Intel oneAPI 4 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
a
October 18 2023
  1 Hour, 33 Minutes
b
October 18 2023
  1 Hour, 40 Minutes
c
October 18 2023
  1 Hour, 41 Minutes
Invert Hiding All Results Option
  1 Hour, 38 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


dddsOpenBenchmarking.orgPhoronix Test SuiteIntel Core i7-1280P @ 4.70GHz (14 Cores / 20 Threads)MSI MS-14C6 (E14C6IMS.115 BIOS)Intel Alder Lake PCH16GB1024GB Micron_3400_MTFDKBA1T0TFHMSI Intel ADL GT2 15GB (1450MHz)Realtek ALC274Intel Alder Lake-P PCH CNVi WiFiUbuntu 23.106.3.0-7-generic (x86_64)GNOME ShellX Server + Wayland4.6 Mesa 23.1.7-1ubuntu1OpenCL 3.0GCC 13.2.0ext41920x1080ProcessorMotherboardChipsetMemoryDiskGraphicsAudioNetworkOSKernelDesktopDisplay ServerOpenGLOpenCLCompilerFile-SystemScreen ResolutionDdds PerformanceSystem Logs- Transparent Huge Pages: madvise- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-13-nEN1TP/gcc-13-13.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-13-nEN1TP/gcc-13-13.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x42c - Thermald 2.5.4 - itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Not affected + tsx_async_abort: Not affected

abcResult OverviewPhoronix Test Suite100%111%122%134%EmbreeIntel Open Image DenoiseOpenVKLoneDNNeasyWaveFluidX3D

dddsonednn: IP Shapes 3D - u8s8f32 - CPUembree: Pathtracer - Asian Dragonembree: Pathtracer - Crownembree: Pathtracer - Asian Dragon Objonednn: IP Shapes 1D - f32 - CPUembree: Pathtracer ISPC - Crownembree: Pathtracer ISPC - Asian Dragononednn: IP Shapes 1D - u8s8f32 - CPUembree: Pathtracer ISPC - Asian Dragon Objopenvkl: vklBenchmarkCPU Scalaroidn: RTLightmap.hdr.4096x4096 - CPU-Onlyonednn: IP Shapes 3D - f32 - CPUoidn: RT.ldr_alb_nrm.3840x2160 - CPU-Onlyopenvkl: vklBenchmarkCPU ISPCoidn: RT.hdr_alb_nrm.3840x2160 - CPU-Onlyeasywave: e2Asean Grid + BengkuluSept2007 Source - 1200onednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUeasywave: e2Asean Grid + BengkuluSept2007 Source - 2400onednn: Deconvolution Batch shapes_1d - f32 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUopenvkl: vklBenchmarkGPU Intel oneAPI SYCLeasywave: e2Asean Grid + BengkuluSept2007 Source - 240onednn: Recurrent Neural Network Inference - u8s8f32 - CPUfluidx3d: FP32-FP16Cfluidx3d: FP32-FP32onednn: Convolution Batch Shapes Auto - f32 - CPUfluidx3d: FP32-FP16Sonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: Deconvolution Batch shapes_3d - f32 - CPUembree: Pathtracer oneAPI SYCL - Crownabc1.714319.48497.52378.568511.45747.703510.1822.276798.8968740.105.638960.211610.22215.082.72839056.794772.43555.28110.05049189.718810.164847.737.7917913710.3224517.556093698.75456463.667729.51053.082097.37445.80016.53789.148045.83827.85613.140496.8812570.086.988250.171320.18240.0972.5721810044.65236.44593.05110.25389853.792015161.737.842811399.9194625.96163658.833586493.682419.55982.250986.27964.98745.68847.778025.3047.19522.494936.8098580.085.851370.171310.19255.6013.014769116.274733.46605.18310.94679946.699513.25159.098.286491459.8034623.656183688.76656443.687889.55125OpenBenchmarking.org

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUcba0.69351.3872.08052.7743.46752.250983.082091.71431MIN: 1.44MIN: 1.45MIN: 1.431. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Embree

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer - Model: Asian Dragoncba36912156.27967.37449.4849MIN: 6.14 / MAX: 12.66MIN: 7.09 / MAX: 12.73MIN: 9.06 / MAX: 12.83

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer - Model: Crowncba2468104.98745.80017.5237MIN: 4.87 / MAX: 10.42MIN: 5.61 / MAX: 10.44MIN: 7.19 / MAX: 10.5

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer - Model: Asian Dragon Objcba2468105.68846.53788.5685MIN: 5.61 / MAX: 5.77MIN: 6.45 / MAX: 7.88MIN: 8.24 / MAX: 11.63

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUcba36912157.778029.1480411.45740MIN: 4.42MIN: 4.54MIN: 4.491. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Embree

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer ISPC - Model: Crowncba2468105.30405.83827.7035MIN: 5.19 / MAX: 7.1MIN: 5.65 / MAX: 10.63MIN: 7.39 / MAX: 10.79

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer ISPC - Model: Asian Dragoncba36912157.19527.856110.1820MIN: 6.95 / MAX: 13.62MIN: 7.45 / MAX: 13.59MIN: 9.65 / MAX: 13.61

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUcba0.70661.41322.11982.82643.5332.494933.140492.27679MIN: 1.7MIN: 1.82MIN: 1.71. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Embree

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer ISPC - Model: Asian Dragon Objcba2468106.80986.88128.8968MIN: 6.54 / MAX: 11.9MIN: 6.52 / MAX: 11.81MIN: 8.47 / MAX: 11.92

OpenVKL

OpenVKL is the Intel Open Volume Kernel Library that offers high-performance volume computation kernels and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems / Sec, More Is BetterOpenVKL 2.0.0Benchmark: vklBenchmarkCPU Scalarcba1632486480585774MIN: 4 / MAX: 1064MIN: 4 / MAX: 1044MIN: 5 / MAX: 1304

Intel Open Image Denoise

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 2.1Run: RTLightmap.hdr.4096x4096 - Device: CPU-Onlycba0.02250.0450.06750.090.11250.080.080.10

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUcba2468105.851376.988255.63896MIN: 5.05MIN: 5.08MIN: 5.041. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Intel Open Image Denoise

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 2.1Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Onlycba0.04730.09460.14190.18920.23650.170.170.21

OpenVKL

OpenVKL is the Intel Open Volume Kernel Library that offers high-performance volume computation kernels and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems / Sec, More Is BetterOpenVKL 2.0.0Benchmark: vklBenchmarkCPU ISPCcba4080120160200131132161MIN: 8 / MAX: 1855MIN: 8 / MAX: 1862MIN: 11 / MAX: 2260

Intel Open Image Denoise

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 2.1Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Onlycba0.04950.0990.14850.1980.24750.190.180.22

easyWave

The easyWave software allows simulating tsunami generation and propagation in the context of early warning systems. EasyWave supports making use of OpenMP for CPU multi-threading and there are also GPU ports available but not currently incorporated as part of this test profile. The easyWave tsunami generation software is run with one of the example/reference input files for measuring the CPU execution time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BettereasyWave r34Input: e2Asean Grid + BengkuluSept2007 Source - Time: 1200cba60120180240300255.60240.10215.081. (CXX) g++ options: -O3 -fopenmp

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUcba0.67831.35662.03492.71323.39153.014762.572182.72830MIN: 2.29MIN: 2.25MIN: 2.221. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUcba2K4K6K8K10K9116.2710044.609056.79MIN: 8924.49MIN: 9728.53MIN: 8880.021. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUcba110022003300440055004733.465236.444772.43MIN: 4556.25MIN: 5033.54MIN: 4570.931. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

easyWave

The easyWave software allows simulating tsunami generation and propagation in the context of early warning systems. EasyWave supports making use of OpenMP for CPU multi-threading and there are also GPU ports available but not currently incorporated as part of this test profile. The easyWave tsunami generation software is run with one of the example/reference input files for measuring the CPU execution time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BettereasyWave r34Input: e2Asean Grid + BengkuluSept2007 Source - Time: 2400cba130260390520650605.18593.05555.281. (CXX) g++ options: -O3 -fopenmp

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPUcba369121510.9510.2510.05MIN: 6.04MIN: 6.02MIN: 6.081. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUcba2K4K6K8K10K9946.699853.709189.71MIN: 9715.98MIN: 9658.57MIN: 8912.941. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUcba2K4K6K8K10K9513.209201.008810.16MIN: 9231.4MIN: 8796.3MIN: 8604.631. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUcba110022003300440055005159.095161.734847.73MIN: 4935.51MIN: 4954.49MIN: 4646.811. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUcba2468108.286497.842817.79179MIN: 7.22MIN: 7.26MIN: 7.191. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenVKL

OpenVKL is the Intel Open Volume Kernel Library that offers high-performance volume computation kernels and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems / Sec, More Is BetterOpenVKL 2.0.0Benchmark: vklBenchmarkGPU Intel oneAPI SYCLcba306090120150145139137MIN: 1 / MAX: 5589MIN: 1 / MAX: 5263MIN: 1 / MAX: 5867

easyWave

The easyWave software allows simulating tsunami generation and propagation in the context of early warning systems. EasyWave supports making use of OpenMP for CPU multi-threading and there are also GPU ports available but not currently incorporated as part of this test profile. The easyWave tsunami generation software is run with one of the example/reference input files for measuring the CPU execution time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BettereasyWave r34Input: e2Asean Grid + BengkuluSept2007 Source - Time: 240cba36912159.8039.91910.3221. (CXX) g++ options: -O3 -fopenmp

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUcba100020003000400050004623.654625.904517.55MIN: 4461.7MIN: 4438.23MIN: 4343.841. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

FluidX3D

FluidX3D is a speedy and memory efficient Boltzmann CFD (Computational Fluid Dynamics) software package implemented using OpenCL and intended for GPU acceleration. FluidX3D is developed by Moritz Lehmann and written free for non-commercial use. This is a test profile measuring the system OpenCL performance using the FluidX3D benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMLUPs/s, More Is BetterFluidX3D 2.9Test: FP32-FP16Ccba130260390520650618616609

OpenBenchmarking.orgMLUPs/s, More Is BetterFluidX3D 2.9Test: FP32-FP32cba80160240320400368365369

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUcba2468108.766508.833588.75450MIN: 7.99MIN: 8.02MIN: 7.931. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

FluidX3D

FluidX3D is a speedy and memory efficient Boltzmann CFD (Computational Fluid Dynamics) software package implemented using OpenCL and intended for GPU acceleration. FluidX3D is developed by Moritz Lehmann and written free for non-commercial use. This is a test profile measuring the system OpenCL performance using the FluidX3D benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMLUPs/s, More Is BetterFluidX3D 2.9Test: FP32-FP16Scba140280420560700644649646

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUcba0.82981.65962.48943.31924.1493.687883.682413.66772MIN: 3.23MIN: 3.24MIN: 3.231. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPUcba36912159.551259.559809.51050MIN: 8.67MIN: 8.72MIN: 8.581. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU

a: The test run did not produce a result.

b: The test run did not produce a result.

c: The test run did not produce a result.

Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU

a: The test run did not produce a result.

b: The test run did not produce a result.

c: The test run did not produce a result.

Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU

a: The test run did not produce a result.

b: The test run did not produce a result.

c: The test run did not produce a result.

Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU

a: The test run did not produce a result.

b: The test run did not produce a result.

c: The test run did not produce a result.

Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU

a: The test run did not produce a result.

b: The test run did not produce a result.

c: The test run did not produce a result.

Embree

Binary: Pathtracer oneAPI SYCL - Model: Asian Dragon Obj

a: The test quit with a non-zero exit status. E: Error: No device of requested type available. -1 (PI_ERROR_DEVICE_NOT_FOUND)

b: The test quit with a non-zero exit status. E: Error: No device of requested type available. -1 (PI_ERROR_DEVICE_NOT_FOUND)

c: The test quit with a non-zero exit status. E: Error: No device of requested type available. -1 (PI_ERROR_DEVICE_NOT_FOUND)

Binary: Pathtracer oneAPI SYCL - Model: Asian Dragon

a: The test quit with a non-zero exit status. E: Error: No device of requested type available. -1 (PI_ERROR_DEVICE_NOT_FOUND)

b: The test quit with a non-zero exit status. E: Error: No device of requested type available. -1 (PI_ERROR_DEVICE_NOT_FOUND)

c: The test quit with a non-zero exit status. E: Error: No device of requested type available. -1 (PI_ERROR_DEVICE_NOT_FOUND)

Binary: Pathtracer oneAPI SYCL - Model: Crown

a: The test quit with a non-zero exit status. E: Error: No device of requested type available. -1 (PI_ERROR_DEVICE_NOT_FOUND)

b: The test quit with a non-zero exit status. E: Error: No device of requested type available. -1 (PI_ERROR_DEVICE_NOT_FOUND)

c: The test quit with a non-zero exit status. E: Error: No device of requested type available. -1 (PI_ERROR_DEVICE_NOT_FOUND)