Xeon E-2278GEL

Intel Xeon E-2278GEL testing with a Logic Supply RXM-181 (Z01-0001A027 BIOS) and Intel HD 3GB on Ubuntu 19.10 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 1910210-HU-XEONE227887
Jump To Table - Results

Statistics

Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
Xeon E-2278GEL
October 20 2019
  6 Hours, 56 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


Xeon E-2278GELOpenBenchmarking.orgPhoronix Test SuiteIntel Xeon E-2278GEL @ 3.90GHz (8 Cores / 16 Threads)Logic Supply RXM-181 (Z01-0001A027 BIOS)Intel Cannon Lake PCH16384MB512GB TS512GMTE510T + 64GB Flash DriveIntel HD 3GB (1150MHz)Realtek ALC233G237HLIntel I219-LM + 2 x Intel I210Ubuntu 19.105.3.0-18-generic (x86_64)GNOME Shell 3.34.1X Server 1.20.5modesetting 1.20.54.5 Mesa 19.2.1GCC 9.2.1 20191008ext41920x1080ProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerDisplay DriverOpenGLCompilerFile-SystemScreen ResolutionXeon E-2278GEL BenchmarksSystem Logs- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-offload-targets=nvptx-none,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - NONE / errors=remount-ro,relatime,rw- Scaling Governor: intel_pstate powersave- OpenJDK Runtime Environment (build 11.0.5-ea+10-post-Ubuntu-0ubuntu1)- l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling

Xeon E-2278GELmkl-dnn: Convolution Batch conv_all - u8s8f32cassandra: Mixed 1:3cassandra: Mixed 1:1cassandra: Readsmkl-dnn: Convolution Batch conv_all - f32mkl-dnn: Deconvolution Batch deconv_all - f32ospray: XFrog Forest - Path Tracerrocksdb: Rand Fill Syncospray: San Miguel - Path Tracermkl-dnn: Convolution Batch conv_googlenet_v3 - f32rocksdb: Rand Fillmkl-dnn: Convolution Batch conv_googlenet_v3 - u8s8f32ospray: San Miguel - SciVisospray: XFrog Forest - SciVismkl-dnn: Convolution Batch conv_3d - u8s8f32ospray: NASA Streamlines - Path Tracerminife: Smallembree: Pathtracer - Asian Dragon Objembree: Pathtracer ISPC - Asian Dragon Objmt-dgemm: Sustained Floating-Point Rateembree: Pathtracer - Crownospray: NASA Streamlines - SciViscassandra: Writesembree: Pathtracer ISPC - Crownrocksdb: Seq Fillembree: Pathtracer - Asian Dragonmkl-dnn: Deconvolution Batch deconv_3d - u8s8f32embree: Pathtracer ISPC - Asian Dragonmkl-dnn: Convolution Batch conv_alexnet - f32mkl-dnn: Convolution Batch conv_3d - f32askap: tConvolve MT - Degriddingaskap: tConvolve MT - Griddingmkl-dnn: IP Batch 1D - f32rocksdb: Read While Writingrocksdb: Rand Readaskap: tConvolve MPI - Degriddingaskap: tConvolve MPI - Griddingmkl-dnn: Deconvolution Batch deconv_1d - u8s8f32mkl-dnn: IP Batch All - u8s8f32mkl-dnn: IP Batch All - f32mkl-dnn: Recurrent Neural Network Training - f32mkl-dnn: Convolution Batch conv_alexnet - u8s8f32mkl-dnn: Deconvolution Batch deconv_1d - f32oidn: Memorialospray: Magnetic Reconnection - SciVisaskap: tConvolve OpenMP - Degriddingaskap: tConvolve OpenMP - Griddingmkl-dnn: IP Batch 1D - u8s8f32ospray: Magnetic Reconnection - Path Tracermkl-dnn: Deconvolution Batch deconv_3d - f32ior: Read Testior: Write TestXeon E-2278GEL56768.302756826914954575.645106.290.9196300.86253.864839632790.1710.211.6716609.772.552534.627.057.642.556.7912.53490667.5810480617.9112970.878.88552.0939.51980.99640.687.91152168731243434980.53645.016855.02330.3351.84421.554404.388.795.129.871121.87683.9051.44141.377.78494.70651.63OpenBenchmarking.org

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_all - Data Type: u8s8f32Xeon E-2278GEL12K24K36K48K60KSE +/- 68.71, N = 356768.30MIN: 55860.51. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

Apache Cassandra

This is a benchmark of the Apache Cassandra NoSQL database management system making use of cassandra-stress. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 3.11.4Test: Mixed 1:3Xeon E-2278GEL6001200180024003000SE +/- 636.62, N = 92756

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 3.11.4Test: Mixed 1:1Xeon E-2278GEL2K4K6K8K10KSE +/- 1093.60, N = 98269

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 3.11.4Test: ReadsXeon E-2278GEL30060090012001500SE +/- 485.50, N = 71495

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_all - Data Type: f32Xeon E-2278GEL10002000300040005000SE +/- 3.08, N = 34575.64MIN: 4314.811. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_all - Data Type: f32Xeon E-2278GEL11002200330044005500SE +/- 6.21, N = 35106.29MIN: 4971.731. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OSPray

Intel OSPray is a portable ray-tracing engine for high-performance, high-fidenlity scientific visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: XFrog Forest - Renderer: Path TracerXeon E-2278GEL0.20480.40960.61440.81921.024SE +/- 0.00, N = 50.91MIN: 0.9

Facebook RocksDB

This is a benchmark of Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.3.6Test: Random Fill SyncXeon E-2278GEL2K4K6K8K10KSE +/- 111.54, N = 1596301. (CXX) g++ options: -O3 -march=native -std=c++11 -fno-builtin-memcmp -fno-rtti -rdynamic -lpthread

OSPray

Intel OSPray is a portable ray-tracing engine for high-performance, high-fidenlity scientific visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: San Miguel - Renderer: Path TracerXeon E-2278GEL0.19350.3870.58050.7740.9675SE +/- 0.00, N = 30.86MIN: 0.81 / MAX: 0.98

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_googlenet_v3 - Data Type: f32Xeon E-2278GEL60120180240300SE +/- 3.01, N = 6253.86MIN: 228.291. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

Facebook RocksDB

This is a benchmark of Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.3.6Test: Random FillXeon E-2278GEL100K200K300K400K500KSE +/- 12423.94, N = 124839631. (CXX) g++ options: -O3 -march=native -std=c++11 -fno-builtin-memcmp -fno-rtti -rdynamic -lpthread

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_googlenet_v3 - Data Type: u8s8f32Xeon E-2278GEL6001200180024003000SE +/- 37.16, N = 52790.17MIN: 2547.411. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OSPray

Intel OSPray is a portable ray-tracing engine for high-performance, high-fidenlity scientific visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: San Miguel - Renderer: SciVisXeon E-2278GEL3691215SE +/- 0.21, N = 1510.21MIN: 9.26 / MAX: 13.33

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: XFrog Forest - Renderer: SciVisXeon E-2278GEL0.37580.75161.12741.50321.879SE +/- 0.00, N = 31.67MIN: 1.65 / MAX: 2.21

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_3d - Data Type: u8s8f32Xeon E-2278GEL4K8K12K16K20KSE +/- 74.35, N = 316609.77MIN: 16421.21. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OSPray

Intel OSPray is a portable ray-tracing engine for high-performance, high-fidenlity scientific visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: NASA Streamlines - Renderer: Path TracerXeon E-2278GEL0.57381.14761.72142.29522.869SE +/- 0.00, N = 42.55MIN: 2.49 / MAX: 3.4

miniFE

MiniFE Finite Element is an application for unstructured implicit finite element codes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgCG Mflops, More Is BetterminiFE 2.2Problem Size: SmallXeon E-2278GEL5001000150020002500SE +/- 0.13, N = 32534.621. (CXX) g++ options: -O3 -fopenmp -pthread -lmpi_cxx -lmpi

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer - Model: Asian Dragon ObjXeon E-2278GEL246810SE +/- 0.01, N = 37.05MIN: 6.86 / MAX: 7.79

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer ISPC - Model: Asian Dragon ObjXeon E-2278GEL246810SE +/- 0.01, N = 37.64MIN: 7.46 / MAX: 8.26

ACES DGEMM

This is a multi-threaded DGEMM benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGFLOP/s, More Is BetterACES DGEMM 1.0Sustained Floating-Point RateXeon E-2278GEL0.57381.14761.72142.29522.869SE +/- 0.01, N = 32.551. (CC) gcc options: -O3 -march=native -fopenmp

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer - Model: CrownXeon E-2278GEL246810SE +/- 0.08, N = 36.79MIN: 6.55 / MAX: 8.2

OSPray

Intel OSPray is a portable ray-tracing engine for high-performance, high-fidenlity scientific visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: NASA Streamlines - Renderer: SciVisXeon E-2278GEL3691215SE +/- 0.26, N = 1412.53MIN: 11.63 / MAX: 16.39

Apache Cassandra

This is a benchmark of the Apache Cassandra NoSQL database management system making use of cassandra-stress. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 3.11.4Test: WritesXeon E-2278GEL11K22K33K44K55KSE +/- 560.62, N = 349066

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer ISPC - Model: CrownXeon E-2278GEL246810SE +/- 0.04, N = 37.58MIN: 7.37 / MAX: 9.07

Facebook RocksDB

This is a benchmark of Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.3.6Test: Sequential FillXeon E-2278GEL200K400K600K800K1000KSE +/- 23602.27, N = 1510480611. (CXX) g++ options: -O3 -march=native -std=c++11 -fno-builtin-memcmp -fno-rtti -rdynamic -lpthread

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer - Model: Asian DragonXeon E-2278GEL246810SE +/- 0.03, N = 37.91MIN: 7.73 / MAX: 8.84

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32Xeon E-2278GEL3K6K9K12K15KSE +/- 2.03, N = 312970.87MIN: 12923.51. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer ISPC - Model: Asian DragonXeon E-2278GEL246810SE +/- 0.03, N = 38.88MIN: 8.72 / MAX: 9.7

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_alexnet - Data Type: f32Xeon E-2278GEL120240360480600SE +/- 5.26, N = 12552.09MIN: 446.631. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_3d - Data Type: f32Xeon E-2278GEL918273645SE +/- 0.49, N = 539.51MIN: 36.691. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

ASKAP

This is a CUDA benchmark of ATNF's ASKAP Benchmark with currently using the tConvolveCuda sub-test. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 2018-11-10Test: tConvolve MT - DegriddingXeon E-2278GEL2004006008001000SE +/- 0.66, N = 3980.991. (CXX) g++ options: -lpthread

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 2018-11-10Test: tConvolve MT - GriddingXeon E-2278GEL140280420560700SE +/- 0.13, N = 3640.681. (CXX) g++ options: -lpthread

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch 1D - Data Type: f32Xeon E-2278GEL246810SE +/- 0.08, N = 127.91MIN: 6.411. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

Facebook RocksDB

This is a benchmark of Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.3.6Test: Read While WritingXeon E-2278GEL300K600K900K1200K1500KSE +/- 9688.78, N = 315216871. (CXX) g++ options: -O3 -march=native -std=c++11 -fno-builtin-memcmp -fno-rtti -rdynamic -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.3.6Test: Random ReadXeon E-2278GEL7M14M21M28M35MSE +/- 519521.07, N = 3312434341. (CXX) g++ options: -O3 -march=native -std=c++11 -fno-builtin-memcmp -fno-rtti -rdynamic -lpthread

ASKAP

This is a CUDA benchmark of ATNF's ASKAP Benchmark with currently using the tConvolveCuda sub-test. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 2018-11-10Test: tConvolve MPI - DegriddingXeon E-2278GEL2004006008001000SE +/- 0.15, N = 3980.531. (CXX) g++ options: -lpthread

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 2018-11-10Test: tConvolve MPI - GriddingXeon E-2278GEL140280420560700SE +/- 0.13, N = 3645.011. (CXX) g++ options: -lpthread

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32Xeon E-2278GEL15003000450060007500SE +/- 81.62, N = 36855.02MIN: 6599.121. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch All - Data Type: u8s8f32Xeon E-2278GEL70140210280350SE +/- 2.99, N = 3330.33MIN: 317.381. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch All - Data Type: f32Xeon E-2278GEL1224364860SE +/- 0.04, N = 351.84MIN: 51.051. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Recurrent Neural Network Training - Data Type: f32Xeon E-2278GEL90180270360450SE +/- 0.53, N = 3421.55MIN: 401.831. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_alexnet - Data Type: u8s8f32Xeon E-2278GEL9001800270036004500SE +/- 49.29, N = 34404.38MIN: 4263.241. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_1d - Data Type: f32Xeon E-2278GEL246810SE +/- 0.13, N = 48.79MIN: 7.631. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

Intel Open Image Denoise

Open Image Denoise is a denoising library for ray-tracing and part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 1.0.0Scene: MemorialXeon E-2278GEL1.1522.3043.4564.6085.76SE +/- 0.07, N = 45.12

OSPray

Intel OSPray is a portable ray-tracing engine for high-performance, high-fidenlity scientific visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: Magnetic Reconnection - Renderer: SciVisXeon E-2278GEL3691215SE +/- 0.03, N = 39.87MIN: 9.52 / MAX: 11.24

ASKAP

This is a CUDA benchmark of ATNF's ASKAP Benchmark with currently using the tConvolveCuda sub-test. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 2018-11-10Test: tConvolve OpenMP - DegriddingXeon E-2278GEL2004006008001000SE +/- 1.57, N = 31121.871. (CXX) g++ options: -lpthread

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 2018-11-10Test: tConvolve OpenMP - GriddingXeon E-2278GEL150300450600750SE +/- 3.11, N = 3683.901. (CXX) g++ options: -lpthread

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch 1D - Data Type: u8s8f32Xeon E-2278GEL1224364860SE +/- 0.17, N = 351.44MIN: 49.871. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OSPray

Intel OSPray is a portable ray-tracing engine for high-performance, high-fidenlity scientific visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: Magnetic Reconnection - Renderer: Path TracerXeon E-2278GEL306090120150SE +/- 1.49, N = 12141.37MIN: 111.11 / MAX: 166.67

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_3d - Data Type: f32Xeon E-2278GEL246810SE +/- 0.01, N = 37.78MIN: 7.731. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

IOR

IOR is a parallel I/O storage benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterIOR 3.2.1Read TestXeon E-2278GEL110220330440550SE +/- 3.15, N = 3494.70MIN: 399.46 / MAX: 560.511. (CC) gcc options: -O2 -lm -pthread -lmpi

OpenBenchmarking.orgMB/s, More Is BetterIOR 3.2.1Write TestXeon E-2278GEL140280420560700SE +/- 7.45, N = 3651.63MIN: 44.97 / MAX: 831.071. (CC) gcc options: -O2 -lm -pthread -lmpi