AWSU20-Multicore Processor: Intel Xeon Platinum 8275CL (4 Cores / 8 Threads), Motherboard: Amazon EC2 c5.2xlarge (1.0 BIOS), Chipset: Intel 440FX 82441FX PMC, Memory: 16384MB, Disk: 215GB Amazon Elastic Block Store, Network: Amazon Elastic
OS: Ubuntu 20.04, Kernel: 5.4.0-1029-aws (x86_64), Compiler: GCC 9.3.0, File-System: ext4, System Layer: KVM
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-9-HskZEa/gcc-9-9.3.0/debian/tmp-nvptx/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vPython Notes: Python 2.7.18 + Python 3.8.5Security Notes: KPTI + usercopy/swapgs barriers and __user pointer sanitization + Full generic retpoline STIBP: disabled RSB filling + PTE Inversion
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 2.0 Encoder Mode: Speed 4 Two-Pass AWSU20-Multicore 0.3533 0.7066 1.0599 1.4132 1.7665 SE +/- 0.00, N = 3 1.57 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 2.0 Encoder Mode: Speed 6 Realtime AWSU20-Multicore 3 6 9 12 15 SE +/- 0.01, N = 3 12.39 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 2.0 Encoder Mode: Speed 6 Two-Pass AWSU20-Multicore 0.5603 1.1206 1.6809 2.2412 2.8015 SE +/- 0.01, N = 3 2.49 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 2.0 Encoder Mode: Speed 8 Realtime AWSU20-Multicore 7 14 21 28 35 SE +/- 0.06, N = 3 28.45 1. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.org Million Grid Points Per Second, More Is Better ASKAP 2018-11-10 Test: tConvolve MT - Degridding AWSU20-Multicore 300 600 900 1200 1500 SE +/- 7.77, N = 3 1358.54 1. (CXX) g++ options: -lpthread
OpenBenchmarking.org Million Grid Points Per Second, More Is Better ASKAP 2018-11-10 Test: tConvolve MPI - Gridding AWSU20-Multicore 400 800 1200 1600 2000 SE +/- 13.33, N = 3 1694.31 1. (CXX) g++ options: -lpthread
OpenBenchmarking.org Million Grid Points Per Second, More Is Better ASKAP 2018-11-10 Test: tConvolve MPI - Degridding AWSU20-Multicore 300 600 900 1200 1500 SE +/- 1.76, N = 3 1367.18 1. (CXX) g++ options: -lpthread
OpenBenchmarking.org Million Grid Points Per Second, More Is Better ASKAP 2018-11-10 Test: tConvolve OpenMP - Gridding AWSU20-Multicore 300 600 900 1200 1500 SE +/- 15.06, N = 3 1521.76 1. (CXX) g++ options: -lpthread
OpenBenchmarking.org Million Grid Points Per Second, More Is Better ASKAP 2018-11-10 Test: tConvolve OpenMP - Degridding AWSU20-Multicore 300 600 900 1200 1500 SE +/- 11.36, N = 3 1444.61 1. (CXX) g++ options: -lpthread
OpenBenchmarking.org Seconds, Fewer Is Better Blender 2.90 Blend File: Classroom - Compute: NVIDIA OptiX AWSU20-Multicore 300 600 900 1200 1500 SE +/- 15.91, N = 3 1413.86
OpenBenchmarking.org Seconds, Fewer Is Better Blender 2.90 Blend File: Fishy Cat - Compute: NVIDIA OptiX AWSU20-Multicore 300 600 900 1200 1500 SE +/- 18.94, N = 3 1204.72
OpenBenchmarking.org Seconds, Fewer Is Better Blender 2.90 Blend File: Barbershop - Compute: NVIDIA OptiX AWSU20-Multicore 400 800 1200 1600 2000 SE +/- 10.04, N = 3 2017.50
OpenBenchmarking.org Seconds, Fewer Is Better Blender 2.90 Blend File: Pabellon Barcelona - Compute: CUDA AWSU20-Multicore 500 1000 1500 2000 2500 SE +/- 6.31, N = 3 2231.00
OpenBenchmarking.org Seconds, Fewer Is Better Blender 2.90 Blend File: Pabellon Barcelona - Compute: OpenCL AWSU20-Multicore 500 1000 1500 2000 2500 SE +/- 9.32, N = 3 2243.32
OpenBenchmarking.org Seconds, Fewer Is Better Blender 2.90 Blend File: Pabellon Barcelona - Compute: CPU-Only AWSU20-Multicore 300 600 900 1200 1500 SE +/- 9.02, N = 3 1606.30
OpenBenchmarking.org Seconds, Fewer Is Better Blender 2.90 Blend File: Pabellon Barcelona - Compute: NVIDIA OptiX AWSU20-Multicore 500 1000 1500 2000 2500 SE +/- 11.58, N = 3 2291.11
C-Ray This is a test of C-Ray, a simple raytracer designed to test the floating-point CPU performance. This test is multi-threaded (16 threads per core), will shoot 8 rays per pixel for anti-aliasing, and will generate a 1600 x 1200 image. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better C-Ray 1.1 Total Time - 4K, 16 Rays Per Pixel AWSU20-Multicore 40 80 120 160 200 SE +/- 0.02, N = 3 178.78 1. (CC) gcc options: -lm -lpthread -O3
OpenBenchmarking.org kH/s - Hash Speed, More Is Better Cpuminer-Opt 3.8.8.1 Algorithm: deep AWSU20-Multicore 800 1600 2400 3200 4000 SE +/- 0.88, N = 3 3657.76 1. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.org kH/s - Hash Speed, More Is Better Cpuminer-Opt 3.8.8.1 Algorithm: lbry AWSU20-Multicore 3K 6K 9K 12K 15K SE +/- 120.14, N = 3 12280 1. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.org kH/s - Hash Speed, More Is Better Cpuminer-Opt 3.8.8.1 Algorithm: skein AWSU20-Multicore 3K 6K 9K 12K 15K SE +/- 24.04, N = 3 14387 1. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.org kH/s - Hash Speed, More Is Better Cpuminer-Opt 3.8.8.1 Algorithm: myr-gr AWSU20-Multicore 1100 2200 3300 4400 5500 SE +/- 5.96, N = 3 5120.76 1. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lssl -lcrypto -lgmp
OpenBenchmarking.org FPS, More Is Better dav1d 0.7.0 Video Input: Summer Nature 4K AWSU20-Multicore 16 32 48 64 80 SE +/- 0.09, N = 3 72.66 1. (CC) gcc options: -pthread
OpenBenchmarking.org FPS, More Is Better dav1d 0.7.0 Video Input: Summer Nature 1080p AWSU20-Multicore 50 100 150 200 250 SE +/- 0.54, N = 3 245.52 1. (CC) gcc options: -pthread
OpenBenchmarking.org FPS, More Is Better dav1d 0.7.0 Video Input: Chimera 1080p 10-bit AWSU20-Multicore 14 28 42 56 70 SE +/- 0.07, N = 3 64.15 1. (CC) gcc options: -pthread
OpenBenchmarking.org Frames Per Second, More Is Better Embree 3.9.0 Binary: Pathtracer ISPC - Model: Crown AWSU20-Multicore 1.0598 2.1196 3.1794 4.2392 5.299 SE +/- 0.06, N = 3 4.71 MIN: 4.53 / MAX: 4.85
OpenBenchmarking.org Frames Per Second, More Is Better Embree 3.9.0 Binary: Pathtracer - Model: Asian Dragon AWSU20-Multicore 1.1205 2.241 3.3615 4.482 5.6025 SE +/- 0.03, N = 3 4.98 MIN: 4.83 / MAX: 5.09
OpenBenchmarking.org Frames Per Second, More Is Better Embree 3.9.0 Binary: Pathtracer - Model: Asian Dragon Obj AWSU20-Multicore 1.0553 2.1106 3.1659 4.2212 5.2765 SE +/- 0.00, N = 3 4.69 MIN: 4.63 / MAX: 4.74
OpenBenchmarking.org Frames Per Second, More Is Better Embree 3.9.0 Binary: Pathtracer ISPC - Model: Asian Dragon AWSU20-Multicore 2 4 6 8 10 SE +/- 0.01, N = 3 6.21 MIN: 6.13 / MAX: 6.33
OpenBenchmarking.org Frames Per Second, More Is Better Embree 3.9.0 Binary: Pathtracer ISPC - Model: Asian Dragon Obj AWSU20-Multicore 1.233 2.466 3.699 4.932 6.165 SE +/- 0.06, N = 7 5.48 MIN: 5.17 / MAX: 5.65
FFmpeg This test uses FFmpeg for testing the system's audio/video encoding performance. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better FFmpeg 4.0.2 H.264 HD To NTSC DV AWSU20-Multicore 2 4 6 8 10 SE +/- 0.06, N = 11 6.69 1. (CC) gcc options: -lavdevice -lavfilter -lavformat -lavcodec -lswresample -lswscale -lavutil -lXv -lX11 -lXext -lm -lxcb -lasound -pthread -lva -lbz2 -llzma -lva-drm -lva-x11 -std=c11 -fomit-frame-pointer -fPIC -O3 -fno-math-errno -fno-signed-zeros -fno-tree-vectorize -MMD -MF -MT
GraphicsMagick This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.33 Operation: Swirl AWSU20-Multicore 30 60 90 120 150 SE +/- 0.33, N = 3 154 1. (CC) gcc options: -fopenmp -O2 -pthread -ljbig -lwebp -lwebpmux -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.33 Operation: Rotate AWSU20-Multicore 90 180 270 360 450 427 1. (CC) gcc options: -fopenmp -O2 -pthread -ljbig -lwebp -lwebpmux -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.33 Operation: Sharpen AWSU20-Multicore 7 14 21 28 35 32 1. (CC) gcc options: -fopenmp -O2 -pthread -ljbig -lwebp -lwebpmux -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.33 Operation: Enhanced AWSU20-Multicore 20 40 60 80 100 80 1. (CC) gcc options: -fopenmp -O2 -pthread -ljbig -lwebp -lwebpmux -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.33 Operation: Resizing AWSU20-Multicore 80 160 240 320 400 SE +/- 1.45, N = 3 371 1. (CC) gcc options: -fopenmp -O2 -pthread -ljbig -lwebp -lwebpmux -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.33 Operation: Noise-Gaussian AWSU20-Multicore 20 40 60 80 100 99 1. (CC) gcc options: -fopenmp -O2 -pthread -ljbig -lwebp -lwebpmux -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.33 Operation: HWB Color Space AWSU20-Multicore 110 220 330 440 550 SE +/- 1.20, N = 3 487 1. (CC) gcc options: -fopenmp -O2 -pthread -ljbig -lwebp -lwebpmux -ltiff -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -llzma -lbz2 -lxml2 -lz -lm -lpthread
OpenBenchmarking.org Average usec, Fewer Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 Exchange AWSU20-Multicore 20 40 60 80 100 SE +/- 1.04, N = 3 81.39 MIN: 78.56 / MAX: 83.5 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
OpenBenchmarking.org Average Mbytes/sec, More Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 PingPong AWSU20-Multicore 1200 2400 3600 4800 6000 SE +/- 42.92, N = 3 5640.12 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
OpenBenchmarking.org Average usec, Fewer Is Better Intel MPI Benchmarks 2019.3 Test: IMB-MPI1 Sendrecv AWSU20-Multicore 12 24 36 48 60 SE +/- 1.06, N = 3 54.65 MIN: 50.65 / MAX: 56.81 1. (CXX) g++ options: -O0 -pedantic -fopenmp -pthread -lmpi_cxx -lmpi
Kvazaar This is a test of Kvazaar as a CPU-based H.265 video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 4K - Video Preset: Slow AWSU20-Multicore 0.4185 0.837 1.2555 1.674 2.0925 SE +/- 0.00, N = 3 1.86 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 4K - Video Preset: Medium AWSU20-Multicore 0.4298 0.8596 1.2894 1.7192 2.149 SE +/- 0.00, N = 3 1.91 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 1080p - Video Preset: Slow AWSU20-Multicore 2 4 6 8 10 SE +/- 0.02, N = 3 8.14 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 1080p - Video Preset: Medium AWSU20-Multicore 2 4 6 8 10 SE +/- 0.00, N = 3 8.39 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 4K - Video Preset: Very Fast AWSU20-Multicore 1.1925 2.385 3.5775 4.77 5.9625 SE +/- 0.01, N = 3 5.30 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 4K - Video Preset: Ultra Fast AWSU20-Multicore 3 6 9 12 15 SE +/- 0.00, N = 3 9.65 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 1080p - Video Preset: Very Fast AWSU20-Multicore 5 10 15 20 25 SE +/- 0.01, N = 3 21.37 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 1080p - Video Preset: Ultra Fast AWSU20-Multicore 9 18 27 36 45 SE +/- 0.06, N = 3 38.34 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O2 -lpthread -lm -lrt
NAMD NAMD is a parallel molecular dynamics code designed for high-performance simulation of large biomolecular systems. NAMD was developed by the Theoretical and Computational Biophysics Group in the Beckman Institute for Advanced Science and Technology at the University of Illinois at Urbana-Champaign. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org days/ns, Fewer Is Better NAMD 2.14 ATPase Simulation - 327,506 Atoms AWSU20-Multicore 1.0198 2.0396 3.0594 4.0792 5.099 SE +/- 0.00338, N = 3 4.53251
NAS Parallel Benchmarks NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: BT.C AWSU20-Multicore 3K 6K 9K 12K 15K SE +/- 45.21, N = 3 14338.01 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 4.0.3
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: CG.C AWSU20-Multicore 1100 2200 3300 4400 5500 SE +/- 16.23, N = 3 5324.40 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 4.0.3
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: EP.C AWSU20-Multicore 110 220 330 440 550 SE +/- 2.64, N = 3 524.41 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 4.0.3
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: EP.D AWSU20-Multicore 110 220 330 440 550 SE +/- 0.68, N = 3 527.15 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 4.0.3
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: FT.C AWSU20-Multicore 2K 4K 6K 8K 10K SE +/- 12.64, N = 3 9494.89 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 4.0.3
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: LU.C AWSU20-Multicore 4K 8K 12K 16K 20K SE +/- 9.22, N = 3 19264.09 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 4.0.3
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: MG.C AWSU20-Multicore 3K 6K 9K 12K 15K SE +/- 25.24, N = 3 14483.73 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 4.0.3
OpenBenchmarking.org Total Mop/s, More Is Better NAS Parallel Benchmarks 3.4 Test / Class: SP.B AWSU20-Multicore 2K 4K 6K 8K 10K SE +/- 12.87, N = 3 10476.46 1. (F9X) gfortran options: -O3 -march=native -pthread -lmpi_usempif08 -lmpi_mpifh -lmpi 2. Open MPI 4.0.3
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch 1D - Data Type: f32 - Engine: CPU AWSU20-Multicore 2 4 6 8 10 SE +/- 0.03, N = 3 8.80 MIN: 8.33 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch All - Data Type: f32 - Engine: CPU AWSU20-Multicore 20 40 60 80 100 SE +/- 0.23, N = 3 105.13 MIN: 102.29 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch 1D - Data Type: u8s8f32 - Engine: CPU AWSU20-Multicore 0.7605 1.521 2.2815 3.042 3.8025 SE +/- 0.02, N = 3 3.38 MIN: 3.19 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch All - Data Type: u8s8f32 - Engine: CPU AWSU20-Multicore 10 20 30 40 50 SE +/- 0.04, N = 3 45.83 MIN: 44.38 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch 1D - Data Type: bf16bf16bf16 - Engine: CPU AWSU20-Multicore 5 10 15 20 25 SE +/- 0.01, N = 3 20.76 MIN: 20.43 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: IP Batch All - Data Type: bf16bf16bf16 - Engine: CPU AWSU20-Multicore 60 120 180 240 300 SE +/- 0.05, N = 3 273.96 MIN: 272.67 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU AWSU20-Multicore 2 4 6 8 10 SE +/- 0.03, N = 3 6.86 MIN: 6.68 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_1d - Data Type: f32 - Engine: CPU AWSU20-Multicore 2 4 6 8 10 SE +/- 0.08, N = 3 7.74 MIN: 7.45 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_3d - Data Type: f32 - Engine: CPU AWSU20-Multicore 3 6 9 12 15 SE +/- 0.01, N = 3 9.37 MIN: 9.23 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU AWSU20-Multicore 2 4 6 8 10 SE +/- 0.02, N = 3 7.23 MIN: 7.05 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32 - Engine: CPU AWSU20-Multicore 0.99 1.98 2.97 3.96 4.95 SE +/- 0.01, N = 3 4.40 MIN: 4.33 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32 - Engine: CPU AWSU20-Multicore 2 4 6 8 10 SE +/- 0.01, N = 3 6.59 MIN: 6.31 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU AWSU20-Multicore 130 260 390 520 650 SE +/- 9.94, N = 3 579.33 MIN: 555.64 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU AWSU20-Multicore 60 120 180 240 300 SE +/- 1.11, N = 3 275.41 MIN: 261.43 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU AWSU20-Multicore 8 16 24 32 40 SE +/- 0.02, N = 3 36.16 MIN: 35.8 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_1d - Data Type: bf16bf16bf16 - Engine: CPU AWSU20-Multicore 9 18 27 36 45 SE +/- 0.02, N = 3 41.47 MIN: 41.27 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Deconvolution Batch deconv_3d - Data Type: bf16bf16bf16 - Engine: CPU AWSU20-Multicore 9 18 27 36 45 SE +/- 0.30, N = 3 38.47 MIN: 38.05 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU AWSU20-Multicore 0.6795 1.359 2.0385 2.718 3.3975 SE +/- 0.01, N = 3 3.02 MIN: 2.87 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU AWSU20-Multicore 0.495 0.99 1.485 1.98 2.475 SE +/- 0.01, N = 3 2.20 MIN: 2.13 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 1.5 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU AWSU20-Multicore 2 4 6 8 10 SE +/- 0.00, N = 3 7.67 MIN: 7.49 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenVINO This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org FPS, More Is Better OpenVINO 2021.1 Model: Face Detection 0106 FP16 - Device: CPU AWSU20-Multicore 0.3308 0.6616 0.9924 1.3232 1.654 SE +/- 0.00, N = 3 1.47 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -pie -pthread -lpthread
OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2021.1 Model: Face Detection 0106 FP16 - Device: CPU AWSU20-Multicore 600 1200 1800 2400 3000 SE +/- 4.13, N = 3 2720.26 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -pie -pthread -lpthread
OpenBenchmarking.org FPS, More Is Better OpenVINO 2021.1 Model: Face Detection 0106 FP32 - Device: CPU AWSU20-Multicore 0.3285 0.657 0.9855 1.314 1.6425 SE +/- 0.01, N = 3 1.46 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -pie -pthread -lpthread
OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2021.1 Model: Face Detection 0106 FP32 - Device: CPU AWSU20-Multicore 600 1200 1800 2400 3000 SE +/- 7.43, N = 3 2727.22 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -pie -pthread -lpthread
OpenBenchmarking.org FPS, More Is Better OpenVINO 2021.1 Model: Person Detection 0106 FP16 - Device: CPU AWSU20-Multicore 0.198 0.396 0.594 0.792 0.99 SE +/- 0.00, N = 3 0.88 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -pie -pthread -lpthread
OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2021.1 Model: Person Detection 0106 FP16 - Device: CPU AWSU20-Multicore 1000 2000 3000 4000 5000 SE +/- 8.04, N = 3 4549.46 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -pie -pthread -lpthread
OpenBenchmarking.org FPS, More Is Better OpenVINO 2021.1 Model: Person Detection 0106 FP32 - Device: CPU AWSU20-Multicore 0.2003 0.4006 0.6009 0.8012 1.0015 SE +/- 0.01, N = 3 0.89 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -pie -pthread -lpthread
OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2021.1 Model: Person Detection 0106 FP32 - Device: CPU AWSU20-Multicore 1000 2000 3000 4000 5000 SE +/- 22.14, N = 3 4459.10 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -pie -pthread -lpthread
OpenBenchmarking.org FPS, More Is Better OpenVINO 2021.1 Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU AWSU20-Multicore 900 1800 2700 3600 4500 SE +/- 2.98, N = 3 4128.98 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -pie -pthread -lpthread
OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2021.1 Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU AWSU20-Multicore 0.2138 0.4276 0.6414 0.8552 1.069 SE +/- 0.00, N = 3 0.95 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -pie -pthread -lpthread
OpenBenchmarking.org FPS, More Is Better OpenVINO 2021.1 Model: Age Gender Recognition Retail 0013 FP32 - Device: CPU AWSU20-Multicore 900 1800 2700 3600 4500 SE +/- 5.14, N = 3 4025.51 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -pie -pthread -lpthread
OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2021.1 Model: Age Gender Recognition Retail 0013 FP32 - Device: CPU AWSU20-Multicore 0.2205 0.441 0.6615 0.882 1.1025 SE +/- 0.00, N = 3 0.98 1. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -pie -pthread -lpthread
OpenBenchmarking.org Items / Sec, More Is Better OpenVKL 0.9 Benchmark: vklBenchmarkStructuredVolume AWSU20-Multicore 6M 12M 18M 24M 30M SE +/- 263669.39, N = 3 30327318.90
OpenBenchmarking.org Items / Sec, More Is Better OpenVKL 0.9 Benchmark: vklBenchmarkUnstructuredVolume AWSU20-Multicore 300K 600K 900K 1200K 1500K SE +/- 1948.30, N = 3 1573006.48
OSPray Intel OSPray is a portable ray-tracing engine for high-performance, high-fidenlity scientific visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org FPS, More Is Better OSPray 1.8.5 Demo: San Miguel - Renderer: SciVis AWSU20-Multicore 2 4 6 8 10 SE +/- 0.00, N = 3 6.06 MIN: 5.95 / MAX: 6.13
OpenBenchmarking.org FPS, More Is Better OSPray 1.8.5 Demo: XFrog Forest - Renderer: SciVis AWSU20-Multicore 0.2228 0.4456 0.6684 0.8912 1.114 SE +/- 0.00, N = 3 0.99 MIN: 0.96 / MAX: 1
OpenBenchmarking.org FPS, More Is Better OSPray 1.8.5 Demo: San Miguel - Renderer: Path Tracer AWSU20-Multicore 0.1215 0.243 0.3645 0.486 0.6075 SE +/- 0.00, N = 3 0.54 MIN: 0.52 / MAX: 0.55
OpenBenchmarking.org FPS, More Is Better OSPray 1.8.5 Demo: XFrog Forest - Renderer: Path Tracer AWSU20-Multicore 0.1238 0.2476 0.3714 0.4952 0.619 SE +/- 0.00, N = 3 0.55 MIN: 0.53
OpenBenchmarking.org FPS, More Is Better OSPray 1.8.5 Demo: Magnetic Reconnection - Renderer: SciVis AWSU20-Multicore 2 4 6 8 10 SE +/- 0.00, N = 3 6.80 MIN: 6.67 / MAX: 6.9
OpenBenchmarking.org FPS, More Is Better OSPray 1.8.5 Demo: NASA Streamlines - Renderer: Path Tracer AWSU20-Multicore 0.3398 0.6796 1.0194 1.3592 1.699 SE +/- 0.00, N = 3 1.51 MIN: 1.44 / MAX: 1.54
Parboil The Parboil Benchmarks from the IMPACT Research Group at University of Illinois are a set of throughput computing applications for looking at computing architecture and compilers. Parboil test-cases support OpenMP, OpenCL, and CUDA multi-processing environments. However, at this time the test profile is just making use of the OpenMP and OpenCL test workloads. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Parboil 2.5 Test: OpenMP LBM AWSU20-Multicore 15 30 45 60 75 SE +/- 0.25, N = 3 66.42 1. (CXX) g++ options: -lm -lpthread -lgomp -O3 -ffast-math -fopenmp
OpenBenchmarking.org Seconds, Fewer Is Better Parboil 2.5 Test: OpenMP CUTCP AWSU20-Multicore 2 4 6 8 10 SE +/- 0.02, N = 3 8.87 1. (CXX) g++ options: -lm -lpthread -lgomp -O3 -ffast-math -fopenmp
OpenBenchmarking.org Seconds, Fewer Is Better Parboil 2.5 Test: OpenMP Stencil AWSU20-Multicore 3 6 9 12 15 SE +/- 0.20, N = 3 12.38 1. (CXX) g++ options: -lm -lpthread -lgomp -O3 -ffast-math -fopenmp
OpenBenchmarking.org Seconds, Fewer Is Better Parboil 2.5 Test: OpenMP MRI Gridding AWSU20-Multicore 11 22 33 44 55 SE +/- 0.28, N = 3 50.38 1. (CXX) g++ options: -lm -lpthread -lgomp -O3 -ffast-math -fopenmp
OpenBenchmarking.org Hydro Cycle Time - Seconds, Fewer Is Better Pennant 1.0.1 Test: leblancbig AWSU20-Multicore 200 400 600 800 1000 SE +/- 0.16, N = 3 826.96 1. (CXX) g++ options: -fopenmp -pthread -lmpi_cxx -lmpi
POV-Ray This is a test of POV-Ray, the Persistence of Vision Raytracer. POV-Ray is used to create 3D graphics using ray-tracing. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better POV-Ray 3.7.0.7 Trace Time AWSU20-Multicore 30 60 90 120 150 SE +/- 0.49, N = 3 129.42 1. (CXX) g++ options: -pipe -O3 -ffast-math -march=native -pthread -lSDL -lXpm -lSM -lICE -lX11 -lIlmImf -lImath -lHalf -lIex -lIexMath -lIlmThread -lpthread -ltiff -ljpeg -lpng -lz -lrt -lm -lboost_thread -lboost_system
Rodinia Rodinia is a suite focused upon accelerating compute-intensive applications with accelerators. CUDA, OpenMP, and OpenCL parallel models are supported by the included applications. This profile utilizes select OpenCL, NVIDIA CUDA and OpenMP test binaries at the moment. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP LavaMD AWSU20-Multicore 120 240 360 480 600 SE +/- 0.07, N = 3 568.63 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP HotSpot3D AWSU20-Multicore 30 60 90 120 150 SE +/- 1.83, N = 3 124.33 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP Leukocyte AWSU20-Multicore 50 100 150 200 250 SE +/- 0.49, N = 3 248.93 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP CFD Solver AWSU20-Multicore 10 20 30 40 50 SE +/- 0.13, N = 3 44.90 1. (CXX) g++ options: -O2 -lOpenCL
OpenBenchmarking.org Seconds, Fewer Is Better Rodinia 3.1 Test: OpenMP Streamcluster AWSU20-Multicore 5 10 15 20 25 SE +/- 0.01, N = 3 22.40 1. (CXX) g++ options: -O2 -lOpenCL
Rust Mandelbrot This test profile is of the combined time for the serial and parallel Mandelbrot sets written in Rustlang via willi-kappler/mandel-rust. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Rust Mandelbrot Time To Complete Serial/Parallel Mandelbrot AWSU20-Multicore 20 40 60 80 100 SE +/- 0.04, N = 3 76.67 1. (CC) gcc options: -m64 -pie -nodefaultlibs -ldl -lrt -lpthread -lgcc_s -lc -lm -lutil
Smallpt Smallpt is a C++ global illumination renderer written in less than 100 lines of code. Global illumination is done via unbiased Monte Carlo path tracing and there is multi-threading support via the OpenMP library. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Smallpt 1.0 Global Illumination Renderer; 128 Samples AWSU20-Multicore 6 12 18 24 30 SE +/- 0.02, N = 3 26.37 1. (CXX) g++ options: -fopenmp -O3
Stockfish This is a test of Stockfish, an advanced C++11 chess benchmark that can scale up to 128 CPU cores. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Nodes Per Second, More Is Better Stockfish 12 Total Time AWSU20-Multicore 2M 4M 6M 8M 10M SE +/- 104055.19, N = 4 7829672 1. (CXX) g++ options: -m64 -lpthread -fno-exceptions -std=c++17 -pedantic -O3 -msse -msse3 -mpopcnt -msse4.1 -mssse3 -msse2 -flto -flto=jobserver
SVT-AV1 This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-AV1 CPU-based multi-threaded video encoder for the AV1 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 0.8 Encoder Mode: Enc Mode 0 - Input: 1080p AWSU20-Multicore 0.0135 0.027 0.0405 0.054 0.0675 SE +/- 0.00, N = 3 0.06 1. (CXX) g++ options: -O3 -fcommon -fPIE -fPIC -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 0.8 Encoder Mode: Enc Mode 4 - Input: 1080p AWSU20-Multicore 0.2903 0.5806 0.8709 1.1612 1.4515 SE +/- 0.01, N = 3 1.29 1. (CXX) g++ options: -O3 -fcommon -fPIE -fPIC -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 0.8 Encoder Mode: Enc Mode 8 - Input: 1080p AWSU20-Multicore 3 6 9 12 15 SE +/- 0.07, N = 3 10.95 1. (CXX) g++ options: -O3 -fcommon -fPIE -fPIC -pie
SVT-HEVC This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.4.1 1080p 8-bit YUV To HEVC Video Encode AWSU20-Multicore 5 10 15 20 25 SE +/- 0.13, N = 3 19.22 1. (CC) gcc options: -fPIE -fPIC -O3 -O2 -pie -rdynamic -lpthread -lrt
SVT-VP9 This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-VP9 CPU-based multi-threaded video encoder for the VP9 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.1 Tuning: VMAF Optimized - Input: Bosphorus 1080p AWSU20-Multicore 20 40 60 80 100 SE +/- 3.15, N = 10 76.13 1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.1 Tuning: PSNR/SSIM Optimized - Input: Bosphorus 1080p AWSU20-Multicore 20 40 60 80 100 SE +/- 0.05, N = 3 80.34 1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.1 Tuning: Visual Quality Optimized - Input: Bosphorus 1080p AWSU20-Multicore 14 28 42 56 70 SE +/- 0.04, N = 3 63.38 1. (CC) gcc options: -O3 -fcommon -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
OpenBenchmarking.org Events Per Second, More Is Better Sysbench 2018-07-28 Test: CPU AWSU20-Multicore 2K 4K 6K 8K 10K SE +/- 1.61, N = 3 7885.38 1. (CC) gcc options: -pthread -O3 -funroll-loops -ggdb3 -march=core2 -rdynamic -ldl -laio -lm
TTSIOD 3D Renderer A portable GPL 3D software renderer that supports OpenMP and Intel Threading Building Blocks with many different rendering modes. This version does not use OpenGL but is entirely CPU/software based. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org FPS, More Is Better TTSIOD 3D Renderer 2.3b Phong Rendering With Soft-Shadow Mapping AWSU20-Multicore 40 80 120 160 200 SE +/- 0.40, N = 3 169.43 1. (CXX) g++ options: -O3 -fomit-frame-pointer -ffast-math -mtune=native -flto -msse -mrecip -mfpmath=sse -msse2 -mssse3 -lSDL -fopenmp -fwhole-program -lstdc++
Tungsten Renderer Tungsten is a C++ physically based renderer that makes use of Intel's Embree ray tracing library. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Tungsten Renderer 0.2.2 Scene: Hair AWSU20-Multicore 15 30 45 60 75 SE +/- 0.06, N = 3 67.68 1. (CXX) g++ options: -std=c++0x -msse2 -msse3 -mssse3 -msse4.1 -msse4.2 -mfma -mbmi2 -mavx512f -mavx512vl -mavx512cd -mavx512dq -mavx512bw -mno-sse4a -mno-avx -mno-avx2 -mno-xop -mno-fma4 -mno-avx512pf -mno-avx512er -mno-avx512ifma -mno-avx512vbmi -fstrict-aliasing -O3 -rdynamic -lIlmImf -lIlmThread -lImath -lHalf -lIex -lz -ljpeg -lpthread -ldl
OpenBenchmarking.org Seconds, Fewer Is Better Tungsten Renderer 0.2.2 Scene: Water Caustic AWSU20-Multicore 10 20 30 40 50 SE +/- 0.13, N = 3 44.77 1. (CXX) g++ options: -std=c++0x -msse2 -msse3 -mssse3 -msse4.1 -msse4.2 -mfma -mbmi2 -mavx512f -mavx512vl -mavx512cd -mavx512dq -mavx512bw -mno-sse4a -mno-avx -mno-avx2 -mno-xop -mno-fma4 -mno-avx512pf -mno-avx512er -mno-avx512ifma -mno-avx512vbmi -fstrict-aliasing -O3 -rdynamic -lIlmImf -lIlmThread -lImath -lHalf -lIex -lz -ljpeg -lpthread -ldl
OpenBenchmarking.org Seconds, Fewer Is Better Tungsten Renderer 0.2.2 Scene: Non-Exponential AWSU20-Multicore 5 10 15 20 25 SE +/- 0.11, N = 3 21.25 1. (CXX) g++ options: -std=c++0x -msse2 -msse3 -mssse3 -msse4.1 -msse4.2 -mfma -mbmi2 -mavx512f -mavx512vl -mavx512cd -mavx512dq -mavx512bw -mno-sse4a -mno-avx -mno-avx2 -mno-xop -mno-fma4 -mno-avx512pf -mno-avx512er -mno-avx512ifma -mno-avx512vbmi -fstrict-aliasing -O3 -rdynamic -lIlmImf -lIlmThread -lImath -lHalf -lIex -lz -ljpeg -lpthread -ldl
OpenBenchmarking.org Seconds, Fewer Is Better Tungsten Renderer 0.2.2 Scene: Volumetric Caustic AWSU20-Multicore 6 12 18 24 30 SE +/- 0.00, N = 3 24.83 1. (CXX) g++ options: -std=c++0x -msse2 -msse3 -mssse3 -msse4.1 -msse4.2 -mfma -mbmi2 -mavx512f -mavx512vl -mavx512cd -mavx512dq -mavx512bw -mno-sse4a -mno-avx -mno-avx2 -mno-xop -mno-fma4 -mno-avx512pf -mno-avx512er -mno-avx512ifma -mno-avx512vbmi -fstrict-aliasing -O3 -rdynamic -lIlmImf -lIlmThread -lImath -lHalf -lIex -lz -ljpeg -lpthread -ldl
x264 This is a simple test of the x264 encoder run on the CPU (OpenCL support disabled) with a sample video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better x264 2019-12-17 H.264 Video Encoding AWSU20-Multicore 8 16 24 32 40 SE +/- 0.31, N = 10 33.67 1. (CC) gcc options: -ldl -lavformat -lavcodec -lavutil -lswscale -m64 -lm -lpthread -O3 -ffast-math -std=gnu99 -fPIC -fomit-frame-pointer -fno-tree-vectorize
x265 This is a simple test of the x265 encoder run on the CPU with 1080p and 4K options for H.265 video encode performance with x265. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better x265 3.4 Video Input: Bosphorus 4K AWSU20-Multicore 1.3298 2.6596 3.9894 5.3192 6.649 SE +/- 0.05, N = 3 5.91 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
OpenBenchmarking.org Frames Per Second, More Is Better x265 3.4 Video Input: Bosphorus 1080p AWSU20-Multicore 7 14 21 28 35 SE +/- 0.34, N = 3 27.89 1. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl -lnuma
AWSU20-Multicore Processor: Intel Xeon Platinum 8275CL (4 Cores / 8 Threads), Motherboard: Amazon EC2 c5.2xlarge (1.0 BIOS), Chipset: Intel 440FX 82441FX PMC, Memory: 16384MB, Disk: 215GB Amazon Elastic Block Store, Network: Amazon Elastic
OS: Ubuntu 20.04, Kernel: 5.4.0-1029-aws (x86_64), Compiler: GCC 9.3.0, File-System: ext4, System Layer: KVM
Compiler Notes: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-9-HskZEa/gcc-9-9.3.0/debian/tmp-nvptx/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -vPython Notes: Python 2.7.18 + Python 3.8.5Security Notes: KPTI + usercopy/swapgs barriers and __user pointer sanitization + Full generic retpoline STIBP: disabled RSB filling + PTE Inversion
Testing initiated at 20 November 2020 07:21 by user root.