Xeon E-2278GEL

Intel Xeon E-2278GEL testing with a Logic Supply RXM-181 (Z01-0001A027 BIOS) and Intel HD 3GB on Ubuntu 19.10 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 1910210-HU-XEONE227887
Jump To Table - Results

Statistics

Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
Xeon E-2278GEL
October 20 2019
  6 Hours, 56 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


Xeon E-2278GELOpenBenchmarking.orgPhoronix Test SuiteIntel Xeon E-2278GEL @ 3.90GHz (8 Cores / 16 Threads)Logic Supply RXM-181 (Z01-0001A027 BIOS)Intel Cannon Lake PCH16384MB512GB TS512GMTE510T + 64GB Flash DriveIntel HD 3GB (1150MHz)Realtek ALC233G237HLIntel I219-LM + 2 x Intel I210Ubuntu 19.105.3.0-18-generic (x86_64)GNOME Shell 3.34.1X Server 1.20.5modesetting 1.20.54.5 Mesa 19.2.1GCC 9.2.1 20191008ext41920x1080ProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerDisplay DriverOpenGLCompilerFile-SystemScreen ResolutionXeon E-2278GEL BenchmarksSystem Logs- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-offload-targets=nvptx-none,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - NONE / errors=remount-ro,relatime,rw- Scaling Governor: intel_pstate powersave- OpenJDK Runtime Environment (build 11.0.5-ea+10-post-Ubuntu-0ubuntu1)- l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling

Xeon E-2278GELmkl-dnn: Deconvolution Batch deconv_1d - u8s8f32mkl-dnn: Convolution Batch conv_all - u8s8f32mkl-dnn: Deconvolution Batch deconv_all - f32mkl-dnn: Deconvolution Batch deconv_3d - f32mkl-dnn: Convolution Batch conv_alexnet - f32mkl-dnn: Convolution Batch conv_alexnet - u8s8f32mkl-dnn: Deconvolution Batch deconv_1d - f32mkl-dnn: Convolution Batch conv_3d - f32mkl-dnn: Convolution Batch conv_all - f32mkl-dnn: IP Batch 1D - u8s8f32mkl-dnn: IP Batch All - u8s8f32mkl-dnn: IP Batch All - f32mkl-dnn: IP Batch 1D - f32askap: tConvolve MT - Griddingmkl-dnn: Deconvolution Batch deconv_3d - u8s8f32mkl-dnn: Convolution Batch conv_3d - u8s8f32mkl-dnn: Recurrent Neural Network Training - f32mkl-dnn: Convolution Batch conv_googlenet_v3 - f32askap: tConvolve MT - Degriddingior: Read Testaskap: tConvolve MPI - Griddingaskap: tConvolve MPI - Degriddingaskap: tConvolve OpenMP - Griddingaskap: tConvolve OpenMP - Degriddingmt-dgemm: Sustained Floating-Point Rateminife: Smallembree: Pathtracer - Crownembree: Pathtracer ISPC - Crownembree: Pathtracer - Asian Dragonembree: Pathtracer - Asian Dragon Objembree: Pathtracer ISPC - Asian Dragonembree: Pathtracer ISPC - Asian Dragon Objoidn: Memorialospray: San Miguel - SciVisospray: XFrog Forest - SciVisospray: San Miguel - Path Tracermkl-dnn: Convolution Batch conv_googlenet_v3 - u8s8f32ospray: NASA Streamlines - SciVisospray: XFrog Forest - Path Tracerospray: Magnetic Reconnection - SciVisospray: NASA Streamlines - Path Tracerospray: Magnetic Reconnection - Path Tracerrocksdb: Rand Fillrocksdb: Rand Readrocksdb: Seq Fillrocksdb: Rand Fill Syncrocksdb: Read While Writingior: Write Testcassandra: Readscassandra: Writescassandra: Mixed 1:1cassandra: Mixed 1:3Xeon E-2278GEL6855.0256768.305106.297.78552.094404.388.7939.514575.6451.44330.3351.847.91640.6812970.8716609.77421.55253.86980.99494.70645.01980.53683.901121.872.552534.626.797.587.917.058.887.645.1210.211.670.862790.1712.530.919.872.55141.3748396331243434104806196301521687651.6314954906682692756OpenBenchmarking.org

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32Xeon E-2278GEL15003000450060007500SE +/- 81.62, N = 36855.02MIN: 6599.121. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_all - Data Type: u8s8f32Xeon E-2278GEL12K24K36K48K60KSE +/- 68.71, N = 356768.30MIN: 55860.51. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_all - Data Type: f32Xeon E-2278GEL11002200330044005500SE +/- 6.21, N = 35106.29MIN: 4971.731. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_3d - Data Type: f32Xeon E-2278GEL246810SE +/- 0.01, N = 37.78MIN: 7.731. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_alexnet - Data Type: f32Xeon E-2278GEL120240360480600SE +/- 5.26, N = 12552.09MIN: 446.631. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_alexnet - Data Type: u8s8f32Xeon E-2278GEL9001800270036004500SE +/- 49.29, N = 34404.38MIN: 4263.241. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_1d - Data Type: f32Xeon E-2278GEL246810SE +/- 0.13, N = 48.79MIN: 7.631. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_3d - Data Type: f32Xeon E-2278GEL918273645SE +/- 0.49, N = 539.51MIN: 36.691. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_all - Data Type: f32Xeon E-2278GEL10002000300040005000SE +/- 3.08, N = 34575.64MIN: 4314.811. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch 1D - Data Type: u8s8f32Xeon E-2278GEL1224364860SE +/- 0.17, N = 351.44MIN: 49.871. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch All - Data Type: u8s8f32Xeon E-2278GEL70140210280350SE +/- 2.99, N = 3330.33MIN: 317.381. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch All - Data Type: f32Xeon E-2278GEL1224364860SE +/- 0.04, N = 351.84MIN: 51.051. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch 1D - Data Type: f32Xeon E-2278GEL246810SE +/- 0.08, N = 127.91MIN: 6.411. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

ASKAP

This is a CUDA benchmark of ATNF's ASKAP Benchmark with currently using the tConvolveCuda sub-test. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 2018-11-10Test: tConvolve MT - GriddingXeon E-2278GEL140280420560700SE +/- 0.13, N = 3640.681. (CXX) g++ options: -lpthread

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32Xeon E-2278GEL3K6K9K12K15KSE +/- 2.03, N = 312970.87MIN: 12923.51. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_3d - Data Type: u8s8f32Xeon E-2278GEL4K8K12K16K20KSE +/- 74.35, N = 316609.77MIN: 16421.21. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Recurrent Neural Network Training - Data Type: f32Xeon E-2278GEL90180270360450SE +/- 0.53, N = 3421.55MIN: 401.831. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_googlenet_v3 - Data Type: f32Xeon E-2278GEL60120180240300SE +/- 3.01, N = 6253.86MIN: 228.291. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

ASKAP

This is a CUDA benchmark of ATNF's ASKAP Benchmark with currently using the tConvolveCuda sub-test. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 2018-11-10Test: tConvolve MT - DegriddingXeon E-2278GEL2004006008001000SE +/- 0.66, N = 3980.991. (CXX) g++ options: -lpthread

IOR

IOR is a parallel I/O storage benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterIOR 3.2.1Read TestXeon E-2278GEL110220330440550SE +/- 3.15, N = 3494.70MIN: 399.46 / MAX: 560.511. (CC) gcc options: -O2 -lm -pthread -lmpi

ASKAP

This is a CUDA benchmark of ATNF's ASKAP Benchmark with currently using the tConvolveCuda sub-test. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 2018-11-10Test: tConvolve MPI - GriddingXeon E-2278GEL140280420560700SE +/- 0.13, N = 3645.011. (CXX) g++ options: -lpthread

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 2018-11-10Test: tConvolve MPI - DegriddingXeon E-2278GEL2004006008001000SE +/- 0.15, N = 3980.531. (CXX) g++ options: -lpthread

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 2018-11-10Test: tConvolve OpenMP - GriddingXeon E-2278GEL150300450600750SE +/- 3.11, N = 3683.901. (CXX) g++ options: -lpthread

OpenBenchmarking.orgMillion Grid Points Per Second, More Is BetterASKAP 2018-11-10Test: tConvolve OpenMP - DegriddingXeon E-2278GEL2004006008001000SE +/- 1.57, N = 31121.871. (CXX) g++ options: -lpthread

ACES DGEMM

This is a multi-threaded DGEMM benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGFLOP/s, More Is BetterACES DGEMM 1.0Sustained Floating-Point RateXeon E-2278GEL0.57381.14761.72142.29522.869SE +/- 0.01, N = 32.551. (CC) gcc options: -O3 -march=native -fopenmp

miniFE

MiniFE Finite Element is an application for unstructured implicit finite element codes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgCG Mflops, More Is BetterminiFE 2.2Problem Size: SmallXeon E-2278GEL5001000150020002500SE +/- 0.13, N = 32534.621. (CXX) g++ options: -O3 -fopenmp -pthread -lmpi_cxx -lmpi

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer - Model: CrownXeon E-2278GEL246810SE +/- 0.08, N = 36.79MIN: 6.55 / MAX: 8.2

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer ISPC - Model: CrownXeon E-2278GEL246810SE +/- 0.04, N = 37.58MIN: 7.37 / MAX: 9.07

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer - Model: Asian DragonXeon E-2278GEL246810SE +/- 0.03, N = 37.91MIN: 7.73 / MAX: 8.84

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer - Model: Asian Dragon ObjXeon E-2278GEL246810SE +/- 0.01, N = 37.05MIN: 6.86 / MAX: 7.79

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer ISPC - Model: Asian DragonXeon E-2278GEL246810SE +/- 0.03, N = 38.88MIN: 8.72 / MAX: 9.7

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer ISPC - Model: Asian Dragon ObjXeon E-2278GEL246810SE +/- 0.01, N = 37.64MIN: 7.46 / MAX: 8.26

Intel Open Image Denoise

Open Image Denoise is a denoising library for ray-tracing and part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 1.0.0Scene: MemorialXeon E-2278GEL1.1522.3043.4564.6085.76SE +/- 0.07, N = 45.12

OSPray

Intel OSPray is a portable ray-tracing engine for high-performance, high-fidenlity scientific visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: San Miguel - Renderer: SciVisXeon E-2278GEL3691215SE +/- 0.21, N = 1510.21MIN: 9.26 / MAX: 13.33

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: XFrog Forest - Renderer: SciVisXeon E-2278GEL0.37580.75161.12741.50321.879SE +/- 0.00, N = 31.67MIN: 1.65 / MAX: 2.21

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: San Miguel - Renderer: Path TracerXeon E-2278GEL0.19350.3870.58050.7740.9675SE +/- 0.00, N = 30.86MIN: 0.81 / MAX: 0.98

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_googlenet_v3 - Data Type: u8s8f32Xeon E-2278GEL6001200180024003000SE +/- 37.16, N = 52790.17MIN: 2547.411. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OSPray

Intel OSPray is a portable ray-tracing engine for high-performance, high-fidenlity scientific visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: NASA Streamlines - Renderer: SciVisXeon E-2278GEL3691215SE +/- 0.26, N = 1412.53MIN: 11.63 / MAX: 16.39

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: XFrog Forest - Renderer: Path TracerXeon E-2278GEL0.20480.40960.61440.81921.024SE +/- 0.00, N = 50.91MIN: 0.9

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: Magnetic Reconnection - Renderer: SciVisXeon E-2278GEL3691215SE +/- 0.03, N = 39.87MIN: 9.52 / MAX: 11.24

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: NASA Streamlines - Renderer: Path TracerXeon E-2278GEL0.57381.14761.72142.29522.869SE +/- 0.00, N = 42.55MIN: 2.49 / MAX: 3.4

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: Magnetic Reconnection - Renderer: Path TracerXeon E-2278GEL306090120150SE +/- 1.49, N = 12141.37MIN: 111.11 / MAX: 166.67

Facebook RocksDB

This is a benchmark of Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.3.6Test: Random FillXeon E-2278GEL100K200K300K400K500KSE +/- 12423.94, N = 124839631. (CXX) g++ options: -O3 -march=native -std=c++11 -fno-builtin-memcmp -fno-rtti -rdynamic -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.3.6Test: Random ReadXeon E-2278GEL7M14M21M28M35MSE +/- 519521.07, N = 3312434341. (CXX) g++ options: -O3 -march=native -std=c++11 -fno-builtin-memcmp -fno-rtti -rdynamic -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.3.6Test: Sequential FillXeon E-2278GEL200K400K600K800K1000KSE +/- 23602.27, N = 1510480611. (CXX) g++ options: -O3 -march=native -std=c++11 -fno-builtin-memcmp -fno-rtti -rdynamic -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.3.6Test: Random Fill SyncXeon E-2278GEL2K4K6K8K10KSE +/- 111.54, N = 1596301. (CXX) g++ options: -O3 -march=native -std=c++11 -fno-builtin-memcmp -fno-rtti -rdynamic -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.3.6Test: Read While WritingXeon E-2278GEL300K600K900K1200K1500KSE +/- 9688.78, N = 315216871. (CXX) g++ options: -O3 -march=native -std=c++11 -fno-builtin-memcmp -fno-rtti -rdynamic -lpthread

IOR

IOR is a parallel I/O storage benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterIOR 3.2.1Write TestXeon E-2278GEL140280420560700SE +/- 7.45, N = 3651.63MIN: 44.97 / MAX: 831.071. (CC) gcc options: -O2 -lm -pthread -lmpi

Apache Cassandra

This is a benchmark of the Apache Cassandra NoSQL database management system making use of cassandra-stress. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 3.11.4Test: ReadsXeon E-2278GEL30060090012001500SE +/- 485.50, N = 71495

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 3.11.4Test: WritesXeon E-2278GEL11K22K33K44K55KSE +/- 560.62, N = 349066

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 3.11.4Test: Mixed 1:1Xeon E-2278GEL2K4K6K8K10KSE +/- 1093.60, N = 98269

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 3.11.4Test: Mixed 1:3Xeon E-2278GEL6001200180024003000SE +/- 636.62, N = 92756