Amazon EC2 c6a.8xlarge

KVM testing on Ubuntu 20.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2202166-NE-2202164NE62
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results

Limit displaying results to tests within:

Timed Code Compilation 2 Tests
CPU Massive 4 Tests
Creator Workloads 3 Tests
HPC - High Performance Computing 7 Tests
Machine Learning 2 Tests
Molecular Dynamics 3 Tests
MPI Benchmarks 2 Tests
Multi-Core 8 Tests
OpenMPI Tests 3 Tests
Programmer / Developer System Benchmarks 3 Tests
Scientific Computing 4 Tests
Server CPU Tests 3 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
c6a.8xlarge
February 16 2022
  1 Hour, 13 Minutes
c5a.8xlarge
February 16 2022
  1 Hour, 2 Minutes
Invert Hiding All Results Option
  1 Hour, 7 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


Amazon EC2 c6a.8xlargeProcessorMotherboardChipsetMemoryDiskNetworkOSKernelVulkanCompilerFile-SystemSystem Layerc6a.8xlargec5a.8xlargeAMD EPYC 7R13 (16 Cores / 32 Threads)Amazon EC2 c6a.8xlarge (1.0 BIOS)Intel 440FX 82441FX PMC62GB107GB Amazon Elastic Block StoreAmazon ElasticUbuntu 20.045.11.0-1022-aws (x86_64)1.1.182GCC 9.3.0ext4KVMAMD EPYC 7R32 (16 Cores / 32 Threads)Amazon EC2 c5a.8xlarge (1.0 BIOS)64GBOpenBenchmarking.orgKernel Details- Transparent Huge Pages: madviseCompiler Details- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-9-HskZEa/gcc-9-9.3.0/debian/tmp-nvptx/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- c6a.8xlarge: CPU Microcode: 0xa001143- c5a.8xlarge: CPU Microcode: 0x8301034Python Details- Python 3.8.10Security Details- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional IBRS_FW STIBP: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected

c6a.8xlarge vs. c5a.8xlarge ComparisonPhoronix Test SuiteBaseline+7.3%+7.3%+14.6%+14.6%+21.9%+21.9%+29.2%+29.2%S.F.P.R29.2%M.R - SciVis26.7%10 - Bosphorus 1080p22.3%NASNet Mobile22.2%MPI CPU - water_GMX50_bare19.8%Inception V419.2%I.R.V19%SqueezeNet18.2%Mobilenet Float17.8%San Miguel - SciVis17.5%1 - Bosphorus 1080p15%ATPase Simulation - 327,506 Atoms13.7%7 - Bosphorus 1080p13.7%NASA Streamlines - SciVis13.6%11.6%Motorbike 30M8.7%Mobilenet Quant8.4%defconfig7.9%Time To Compile7.4%ACES DGEMMOSPraySVT-HEVCTensorFlow LiteGROMACSTensorFlow LiteTensorFlow LiteTensorFlow LiteTensorFlow LiteOSPraySVT-HEVCNAMDSVT-HEVCOSPrayHigh Performance Conjugate GradientOpenFOAMTensorFlow LiteTimed Linux Kernel CompilationTimed Godot Game Engine Compilationc6a.8xlargec5a.8xlarge

Amazon EC2 c6a.8xlargehpcg: namd: ATPase Simulation - 327,506 Atomsopenfoam: Motorbike 30Mospray: San Miguel - SciVisospray: NASA Streamlines - SciVisospray: Magnetic Reconnection - SciVissvt-hevc: 1 - Bosphorus 1080psvt-hevc: 7 - Bosphorus 1080psvt-hevc: 10 - Bosphorus 1080pmt-dgemm: Sustained Floating-Point Ratebuild-godot: Time To Compilebuild-linux-kernel: defconfiggromacs: MPI CPU - water_GMX50_baretensorflow-lite: SqueezeNettensorflow-lite: Inception V4tensorflow-lite: NASNet Mobiletensorflow-lite: Mobilenet Floattensorflow-lite: Mobilenet Quanttensorflow-lite: Inception ResNet V2tnn: CPU - DenseNettnn: CPU - MobileNet v2tnn: CPU - SqueezeNet v2tnn: CPU - SqueezeNet v1.1c6a.8xlargec5a.8xlarge10.40731.2480367.652534.1018.1814.22178.33366.385.15979395.09362.4372.032108293153059712121871160.981238.113689973257.110318.29078.997282.6959.328051.4194073.5321.2830.0114.3512.37156.80299.603.992451102.17367.3591.696128039182519714809283805.288088.71628750OpenBenchmarking.org

High Performance Conjugate Gradient

HPCG is the High Performance Conjugate Gradient and is a new scientific benchmark from Sandia National Lans focused for super-computer testing with modern real-world workloads compared to HPCC. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGFLOP/s, More Is BetterHigh Performance Conjugate Gradient 3.1c6a.8xlargec5a.8xlarge3691215SE +/- 0.07295, N = 3SE +/- 0.03245, N = 310.407309.328051. (CXX) g++ options: -O3 -ffast-math -ftree-vectorize -pthread -lmpi_cxx -lmpi
OpenBenchmarking.orgGFLOP/s, More Is BetterHigh Performance Conjugate Gradient 3.1c6a.8xlargec5a.8xlarge3691215Min: 10.26 / Avg: 10.41 / Max: 10.48Min: 9.29 / Avg: 9.33 / Max: 9.391. (CXX) g++ options: -O3 -ffast-math -ftree-vectorize -pthread -lmpi_cxx -lmpi

NAMD

NAMD is a parallel molecular dynamics code designed for high-performance simulation of large biomolecular systems. NAMD was developed by the Theoretical and Computational Biophysics Group in the Beckman Institute for Advanced Science and Technology at the University of Illinois at Urbana-Champaign. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgdays/ns, Fewer Is BetterNAMD 2.14ATPase Simulation - 327,506 Atomsc6a.8xlargec5a.8xlarge0.31940.63880.95821.27761.597SE +/- 0.01177, N = 7SE +/- 0.00050, N = 31.248031.41940
OpenBenchmarking.orgdays/ns, Fewer Is BetterNAMD 2.14ATPase Simulation - 327,506 Atomsc6a.8xlargec5a.8xlarge246810Min: 1.23 / Avg: 1.25 / Max: 1.31Min: 1.42 / Avg: 1.42 / Max: 1.42

OpenFOAM

OpenFOAM is the leading free, open source software for computational fluid dynamics (CFD). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenFOAM 8Input: Motorbike 30Mc6a.8xlargec5a.8xlarge1632486480SE +/- 0.09, N = 3SE +/- 0.13, N = 367.6573.531. (CXX) g++ options: -std=c++11 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm
OpenBenchmarking.orgSeconds, Fewer Is BetterOpenFOAM 8Input: Motorbike 30Mc6a.8xlargec5a.8xlarge1428425670Min: 67.52 / Avg: 67.65 / Max: 67.83Min: 73.29 / Avg: 73.53 / Max: 73.731. (CXX) g++ options: -std=c++11 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm

OSPray

Intel OSPray is a portable ray-tracing engine for high-performance, high-fidenlity scientific visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: San Miguel - Renderer: SciVisc6a.8xlargec5a.8xlarge612182430SE +/- 0.00, N = 325.0021.28MIN: 17.86 / MAX: 27.03MIN: 20.41 / MAX: 22.73
OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: San Miguel - Renderer: SciVisc6a.8xlargec5a.8xlarge612182430Min: 21.28 / Avg: 21.28 / Max: 21.28

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: NASA Streamlines - Renderer: SciVisc6a.8xlargec5a.8xlarge816243240SE +/- 0.38, N = 3SE +/- 0.30, N = 334.1030.01MIN: 13.7 / MAX: 34.48MIN: 29.41 / MAX: 30.3
OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: NASA Streamlines - Renderer: SciVisc6a.8xlargec5a.8xlarge714212835Min: 33.33 / Avg: 34.1 / Max: 34.48Min: 29.41 / Avg: 30.01 / Max: 30.3

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: Magnetic Reconnection - Renderer: SciVisc6a.8xlargec5a.8xlarge48121620SE +/- 0.00, N = 3SE +/- 0.07, N = 318.1814.35MIN: 8.26 / MAX: 18.52MIN: 14.08 / MAX: 14.49
OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: Magnetic Reconnection - Renderer: SciVisc6a.8xlargec5a.8xlarge510152025Min: 18.18 / Avg: 18.18 / Max: 18.18Min: 14.29 / Avg: 14.35 / Max: 14.49

SVT-HEVC

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 1 - Input: Bosphorus 1080pc6a.8xlargec5a.8xlarge48121620SE +/- 0.02, N = 3SE +/- 0.01, N = 314.2212.371. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 1 - Input: Bosphorus 1080pc6a.8xlargec5a.8xlarge48121620Min: 14.18 / Avg: 14.22 / Max: 14.26Min: 12.36 / Avg: 12.37 / Max: 12.381. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 7 - Input: Bosphorus 1080pc6a.8xlargec5a.8xlarge4080120160200SE +/- 0.71, N = 3SE +/- 0.28, N = 3178.33156.801. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 7 - Input: Bosphorus 1080pc6a.8xlargec5a.8xlarge306090120150Min: 177.51 / Avg: 178.33 / Max: 179.75Min: 156.37 / Avg: 156.8 / Max: 157.321. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 10 - Input: Bosphorus 1080pc6a.8xlargec5a.8xlarge80160240320400SE +/- 0.66, N = 3SE +/- 0.61, N = 3366.38299.601. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-HEVC 1.5.0Tuning: 10 - Input: Bosphorus 1080pc6a.8xlargec5a.8xlarge70140210280350Min: 365.41 / Avg: 366.38 / Max: 367.65Min: 298.66 / Avg: 299.6 / Max: 300.751. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt

ACES DGEMM

This is a multi-threaded DGEMM benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGFLOP/s, More Is BetterACES DGEMM 1.0Sustained Floating-Point Ratec6a.8xlargec5a.8xlarge1.1612.3223.4834.6445.805SE +/- 0.025529, N = 3SE +/- 0.021705, N = 35.1597933.9924511. (CC) gcc options: -O3 -march=native -fopenmp
OpenBenchmarking.orgGFLOP/s, More Is BetterACES DGEMM 1.0Sustained Floating-Point Ratec6a.8xlargec5a.8xlarge246810Min: 5.11 / Avg: 5.16 / Max: 5.19Min: 3.95 / Avg: 3.99 / Max: 4.031. (CC) gcc options: -O3 -march=native -fopenmp

Timed Godot Game Engine Compilation

This test times how long it takes to compile the Godot Game Engine. Godot is a popular, open-source, cross-platform 2D/3D game engine and is built using the SCons build system and targeting the X11 platform. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Godot Game Engine Compilation 3.2.3Time To Compilec6a.8xlargec5a.8xlarge20406080100SE +/- 0.22, N = 3SE +/- 0.13, N = 395.09102.17
OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Godot Game Engine Compilation 3.2.3Time To Compilec6a.8xlargec5a.8xlarge20406080100Min: 94.82 / Avg: 95.09 / Max: 95.52Min: 101.93 / Avg: 102.17 / Max: 102.36

Timed Linux Kernel Compilation

This test times how long it takes to build the Linux kernel in a default configuration (defconfig) for the architecture being tested or alternatively an allmodconfig for building all possible kernel modules for the build. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 5.16Build: defconfigc6a.8xlargec5a.8xlarge1530456075SE +/- 0.22, N = 3SE +/- 0.10, N = 362.4467.36
OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 5.16Build: defconfigc6a.8xlargec5a.8xlarge1326395265Min: 62.03 / Avg: 62.44 / Max: 62.8Min: 67.17 / Avg: 67.36 / Max: 67.51

GROMACS

The GROMACS (GROningen MAchine for Chemical Simulations) molecular dynamics package testing with the water_GMX50 data. This test profile allows selecting between CPU and GPU-based GROMACS builds. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgNs Per Day, More Is BetterGROMACS 2021.2Implementation: MPI CPU - Input: water_GMX50_barec6a.8xlargec5a.8xlarge0.45720.91441.37161.82882.286SE +/- 0.002, N = 3SE +/- 0.000, N = 32.0321.6961. (CXX) g++ options: -O3 -pthread
OpenBenchmarking.orgNs Per Day, More Is BetterGROMACS 2021.2Implementation: MPI CPU - Input: water_GMX50_barec6a.8xlargec5a.8xlarge246810Min: 2.03 / Avg: 2.03 / Max: 2.04Min: 1.7 / Avg: 1.7 / Max: 1.71. (CXX) g++ options: -O3 -pthread

TensorFlow Lite

This is a benchmark of the TensorFlow Lite implementation. The current Linux support is limited to running on CPUs. This test profile is measuring the average inference time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: SqueezeNetc6a.8xlargec5a.8xlarge30K60K90K120K150KSE +/- 203.21, N = 3SE +/- 241.62, N = 3108293128039
OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: SqueezeNetc6a.8xlargec5a.8xlarge20K40K60K80K100KMin: 107896 / Avg: 108293.33 / Max: 108566Min: 127703 / Avg: 128039.33 / Max: 128508

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Inception V4c6a.8xlargec5a.8xlarge400K800K1200K1600K2000KSE +/- 1506.68, N = 3SE +/- 3243.08, N = 315305971825197
OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Inception V4c6a.8xlargec5a.8xlarge300K600K900K1200K1500KMin: 1529080 / Avg: 1530596.67 / Max: 1533610Min: 1818740 / Avg: 1825196.67 / Max: 1828960

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: NASNet Mobilec6a.8xlargec5a.8xlarge30K60K90K120K150KSE +/- 1436.57, N = 3SE +/- 138.67, N = 3121218148092
OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: NASNet Mobilec6a.8xlargec5a.8xlarge30K60K90K120K150KMin: 118346 / Avg: 121218 / Max: 122724Min: 147835 / Avg: 148091.67 / Max: 148311

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Mobilenet Floatc6a.8xlargec5a.8xlarge20K40K60K80K100KSE +/- 210.72, N = 3SE +/- 381.99, N = 371160.983805.2
OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Mobilenet Floatc6a.8xlargec5a.8xlarge15K30K45K60K75KMin: 70749 / Avg: 71160.93 / Max: 71444Min: 83067.6 / Avg: 83805.23 / Max: 84346.3

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Mobilenet Quantc6a.8xlargec5a.8xlarge20K40K60K80K100KSE +/- 217.08, N = 3SE +/- 147.81, N = 381238.188088.7
OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Mobilenet Quantc6a.8xlargec5a.8xlarge15K30K45K60K75KMin: 80912 / Avg: 81238.13 / Max: 81649.4Min: 87882.9 / Avg: 88088.73 / Max: 88375.4

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Inception ResNet V2c6a.8xlargec5a.8xlarge300K600K900K1200K1500KSE +/- 398.43, N = 3SE +/- 6942.31, N = 313689971628750
OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2020-08-23Model: Inception ResNet V2c6a.8xlargec5a.8xlarge300K600K900K1200K1500KMin: 1368200 / Avg: 1368996.67 / Max: 1369410Min: 1620600 / Avg: 1628750 / Max: 1642560

TNN

TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: DenseNetc6a.8xlarge7001400210028003500SE +/- 2.96, N = 33257.11MIN: 3092.98 / MAX: 3850.221. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: MobileNet v2c6a.8xlarge70140210280350SE +/- 4.12, N = 3318.29MIN: 308.92 / MAX: 958.391. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: SqueezeNet v2c6a.8xlarge20406080100SE +/- 0.29, N = 379.00MIN: 78.5 / MAX: 117.411. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: SqueezeNet v1.1c6a.8xlarge60120180240300SE +/- 0.23, N = 3282.70MIN: 280.75 / MAX: 353.261. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl