EPYC 7601 Extra

AMD EPYC 7601 32-Core testing with a ASRockRack EPYCD8 (P2.10 BIOS) and llvmpipe 126GB on Ubuntu 19.10 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 1911029-HU-EPYC7601E78
Jump To Table - Results

Statistics

Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
EPYC 7601
November 01 2019
  4 Hours, 37 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


EPYC 7601 ExtraOpenBenchmarking.orgPhoronix Test SuiteAMD EPYC 7601 32-Core @ 2.20GHz (32 Cores / 64 Threads)ASRockRack EPYCD8 (P2.10 BIOS)AMD 17h129024MB280GB INTEL SSDPED1D280GAllvmpipe 126GB2 x Intel I350Ubuntu 19.105.4.0-999-generic (x86_64) 20191030GNOME Shell 3.34.1X Server 1.20.5modesetting 1.20.53.3 Mesa 19.2.1 (LLVM 9.0 128 bits)GCC 9.2.1 20191008ext41024x768ProcessorMotherboardChipsetMemoryDiskGraphicsNetworkOSKernelDesktopDisplay ServerDisplay DriverOpenGLCompilerFile-SystemScreen ResolutionEPYC 7601 Extra BenchmarksSystem Logs- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-offload-targets=nvptx-none,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: acpi-cpufreq ondemand- l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional STIBP: disabled RSB filling

EPYC 7601 Extrafftw: Stock - 1D FFT Size 32fftw: Stock - 1D FFT Size 64fftw: Stock - 2D FFT Size 32fftw: Stock - 2D FFT Size 64fftw: Stock - 1D FFT Size 128fftw: Stock - 1D FFT Size 256fftw: Stock - 1D FFT Size 512fftw: Stock - 2D FFT Size 128fftw: Stock - 2D FFT Size 256fftw: Stock - 2D FFT Size 512fftw: Stock - 1D FFT Size 1024fftw: Stock - 1D FFT Size 2048fftw: Stock - 1D FFT Size 4096fftw: Stock - 2D FFT Size 1024fftw: Stock - 2D FFT Size 2048fftw: Stock - 2D FFT Size 4096fftw: Float + SSE - 1D FFT Size 32fftw: Float + SSE - 1D FFT Size 64fftw: Float + SSE - 2D FFT Size 32fftw: Float + SSE - 2D FFT Size 64fftw: Float + SSE - 1D FFT Size 128fftw: Float + SSE - 1D FFT Size 256fftw: Float + SSE - 1D FFT Size 512fftw: Float + SSE - 2D FFT Size 128fftw: Float + SSE - 2D FFT Size 256fftw: Float + SSE - 2D FFT Size 512fftw: Float + SSE - 1D FFT Size 1024fftw: Float + SSE - 1D FFT Size 2048fftw: Float + SSE - 1D FFT Size 4096fftw: Float + SSE - 2D FFT Size 1024fftw: Float + SSE - 2D FFT Size 2048fftw: Float + SSE - 2D FFT Size 4096mkl-dnn: IP Batch 1D - f32mkl-dnn: IP Batch All - f32mkl-dnn: IP Batch 1D - u8s8f32mkl-dnn: IP Batch All - u8s8f32mkl-dnn: Convolution Batch conv_3d - f32mkl-dnn: Convolution Batch conv_all - f32mkl-dnn: Convolution Batch conv_3d - u8s8f32mkl-dnn: Deconvolution Batch deconv_1d - f32mkl-dnn: Deconvolution Batch deconv_3d - f32mkl-dnn: Convolution Batch conv_alexnet - f32mkl-dnn: Convolution Batch conv_all - u8s8f32mkl-dnn: Deconvolution Batch deconv_all - f32mkl-dnn: Deconvolution Batch deconv_1d - u8s8f32mkl-dnn: Deconvolution Batch deconv_3d - u8s8f32mkl-dnn: Recurrent Neural Network Training - f32mkl-dnn: Convolution Batch conv_alexnet - u8s8f32mkl-dnn: Convolution Batch conv_googlenet_v3 - f32mkl-dnn: Convolution Batch conv_googlenet_v3 - u8s8f32dav1d: Chimera 1080pdav1d: Summer Nature 4Kdav1d: Summer Nature 1080pdav1d: Chimera 1080p 10-bitospray: San Miguel - SciVisospray: XFrog Forest - SciVisospray: San Miguel - Path Tracerospray: NASA Streamlines - SciVisospray: XFrog Forest - Path Tracerospray: Magnetic Reconnection - SciVisospray: NASA Streamlines - Path Tracerospray: Magnetic Reconnection - Path Tracerembree: Pathtracer - Crownembree: Pathtracer ISPC - Crownembree: Pathtracer - Asian Dragonembree: Pathtracer - Asian Dragon Objembree: Pathtracer ISPC - Asian Dragonembree: Pathtracer ISPC - Asian Dragon Objsvt-av1: Enc Mode 0 - 1080psvt-av1: Enc Mode 4 - 1080psvt-av1: Enc Mode 8 - 1080poidn: MemorialEPYC 76016306.05728.38158.46452.56006.76420.56481.36329.06141.86114.16610.26388.66316.35521.05065.24595.2106361334126308261251640120976256632549223334235762772028028277422088914876134234.8891929.494553.6097272.37618.67102566.444899.466.522088.74631303.91722469.13129.281877.273138.11413.2362480.57147.5701383.22314.98151.58356.3361.5524.394.341.9532.262.2818.126.3720020.949219.544124.125121.591523.114920.12820.0424.08940.1917.15OpenBenchmarking.org

FFTW

FFTW is a C subroutine library for computing the discrete Fourier transform (DFT) in one or more dimensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 1D FFT Size 32EPYC 760114002800420056007000SE +/- 11.62, N = 36306.01. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 1D FFT Size 64EPYC 760112002400360048006000SE +/- 3.36, N = 35728.31. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 2D FFT Size 32EPYC 76012K4K6K8K10KSE +/- 1.38, N = 38158.41. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 2D FFT Size 64EPYC 760114002800420056007000SE +/- 6.55, N = 36452.51. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 1D FFT Size 128EPYC 760113002600390052006500SE +/- 3.46, N = 36006.71. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 1D FFT Size 256EPYC 760114002800420056007000SE +/- 4.29, N = 36420.51. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 1D FFT Size 512EPYC 760114002800420056007000SE +/- 4.01, N = 36481.31. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 2D FFT Size 128EPYC 760114002800420056007000SE +/- 38.57, N = 36329.01. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 2D FFT Size 256EPYC 760113002600390052006500SE +/- 5.21, N = 36141.81. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 2D FFT Size 512EPYC 760113002600390052006500SE +/- 12.86, N = 36114.11. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 1D FFT Size 1024EPYC 760114002800420056007000SE +/- 4.09, N = 36610.21. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 1D FFT Size 2048EPYC 760114002800420056007000SE +/- 5.48, N = 36388.61. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 1D FFT Size 4096EPYC 760114002800420056007000SE +/- 5.98, N = 36316.31. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 2D FFT Size 1024EPYC 760112002400360048006000SE +/- 6.38, N = 35521.01. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 2D FFT Size 2048EPYC 760111002200330044005500SE +/- 25.21, N = 35065.21. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Stock - Size: 2D FFT Size 4096EPYC 760110002000300040005000SE +/- 4.25, N = 34595.21. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 1D FFT Size 32EPYC 76012K4K6K8K10KSE +/- 8.33, N = 3106361. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 1D FFT Size 64EPYC 76013K6K9K12K15KSE +/- 57.19, N = 3133411. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 2D FFT Size 32EPYC 76016K12K18K24K30KSE +/- 12.17, N = 3263081. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 2D FFT Size 64EPYC 76016K12K18K24K30KSE +/- 32.17, N = 3261251. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 1D FFT Size 128EPYC 76014K8K12K16K20KSE +/- 27.94, N = 3164011. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 1D FFT Size 256EPYC 76014K8K12K16K20KSE +/- 137.11, N = 3209761. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 1D FFT Size 512EPYC 76015K10K15K20K25KSE +/- 85.41, N = 3256631. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 2D FFT Size 128EPYC 76015K10K15K20K25KSE +/- 42.81, N = 3254921. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 2D FFT Size 256EPYC 76015K10K15K20K25KSE +/- 45.35, N = 3233341. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 2D FFT Size 512EPYC 76015K10K15K20K25KSE +/- 25.75, N = 3235761. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 1D FFT Size 1024EPYC 76016K12K18K24K30KSE +/- 232.71, N = 3277201. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 1D FFT Size 2048EPYC 76016K12K18K24K30KSE +/- 284.85, N = 3280281. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 1D FFT Size 4096EPYC 76016K12K18K24K30KSE +/- 42.27, N = 3277421. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 2D FFT Size 1024EPYC 76014K8K12K16K20KSE +/- 47.22, N = 3208891. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 2D FFT Size 2048EPYC 76013K6K9K12K15KSE +/- 29.63, N = 3148761. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

OpenBenchmarking.orgMflops, More Is BetterFFTW 3.3.6Build: Float + SSE - Size: 2D FFT Size 4096EPYC 76013K6K9K12K15KSE +/- 46.89, N = 3134231. (CC) gcc options: -pthread -O3 -fomit-frame-pointer -mtune=native -malign-double -fstrict-aliasing -fno-schedule-insns -ffast-math -lm

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch 1D - Data Type: f32EPYC 76011.10012.20023.30034.40045.5005SE +/- 0.07959, N = 154.88919MIN: 3.221. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch All - Data Type: f32EPYC 7601714212835SE +/- 0.43, N = 329.49MIN: 27.71. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch 1D - Data Type: u8s8f32EPYC 76011224364860SE +/- 0.18, N = 353.61MIN: 46.591. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch All - Data Type: u8s8f32EPYC 760160120180240300SE +/- 0.40, N = 3272.38MIN: 258.781. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_3d - Data Type: f32EPYC 7601510152025SE +/- 0.08, N = 318.67MIN: 17.411. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_all - Data Type: f32EPYC 76016001200180024003000SE +/- 2.32, N = 32566.44MIN: 2437.381. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_3d - Data Type: u8s8f32EPYC 760111002200330044005500SE +/- 3.69, N = 34899.46MIN: 4872.71. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_1d - Data Type: f32EPYC 7601246810SE +/- 0.07721, N = 66.52208MIN: 5.591. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_3d - Data Type: f32EPYC 7601246810SE +/- 0.11045, N = 158.74631MIN: 6.951. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_alexnet - Data Type: f32EPYC 760170140210280350SE +/- 1.22, N = 3303.92MIN: 291.881. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_all - Data Type: u8s8f32EPYC 76015K10K15K20K25KSE +/- 59.53, N = 322469.1MIN: 21801.71. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_all - Data Type: f32EPYC 76017001400210028003500SE +/- 23.55, N = 33129.28MIN: 2970.161. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f32EPYC 7601400800120016002000SE +/- 9.10, N = 31877.27MIN: 1855.951. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f32EPYC 76017001400210028003500SE +/- 5.06, N = 33138.11MIN: 3124.281. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Recurrent Neural Network Training - Data Type: f32EPYC 760190180270360450SE +/- 1.62, N = 3413.24MIN: 395.511. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_alexnet - Data Type: u8s8f32EPYC 76015001000150020002500SE +/- 5.12, N = 32480.57MIN: 2430.351. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_googlenet_v3 - Data Type: f32EPYC 7601306090120150SE +/- 1.38, N = 3147.57MIN: 128.591. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_googlenet_v3 - Data Type: u8s8f32EPYC 760130060090012001500SE +/- 3.82, N = 31383.22MIN: 1332.561. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -ldl

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.5.0Video Input: Chimera 1080pEPYC 760170140210280350SE +/- 0.84, N = 3314.98MIN: 247.02 / MAX: 383.841. (CC) gcc options: -pthread

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.5.0Video Input: Summer Nature 4KEPYC 7601306090120150SE +/- 0.45, N = 3151.58MIN: 94.31 / MAX: 164.181. (CC) gcc options: -pthread

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.5.0Video Input: Summer Nature 1080pEPYC 760180160240320400SE +/- 1.25, N = 3356.33MIN: 203.6 / MAX: 392.891. (CC) gcc options: -pthread

OpenBenchmarking.orgFPS, More Is Betterdav1d 0.5.0Video Input: Chimera 1080p 10-bitEPYC 76011428425670SE +/- 0.15, N = 361.55MIN: 39.53 / MAX: 110.871. (CC) gcc options: -pthread

OSPray

Intel OSPray is a portable ray-tracing engine for high-performance, high-fidenlity scientific visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: San Miguel - Renderer: SciVisEPYC 7601612182430SE +/- 0.00, N = 1224.39MIN: 14.93 / MAX: 26.32

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: XFrog Forest - Renderer: SciVisEPYC 76010.97651.9532.92953.9064.8825SE +/- 0.01, N = 34.34MIN: 3.95 / MAX: 4.41

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: San Miguel - Renderer: Path TracerEPYC 76010.43880.87761.31641.75522.194SE +/- 0.00, N = 31.95MIN: 1.54 / MAX: 1.98

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: NASA Streamlines - Renderer: SciVisEPYC 7601714212835SE +/- 0.00, N = 1232.26MIN: 12.82 / MAX: 33.33

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: XFrog Forest - Renderer: Path TracerEPYC 76010.5131.0261.5392.0522.565SE +/- 0.00, N = 32.28MIN: 2.11 / MAX: 2.31

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: Magnetic Reconnection - Renderer: SciVisEPYC 760148121620SE +/- 0.06, N = 518.12MIN: 9.17 / MAX: 18.52

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: NASA Streamlines - Renderer: Path TracerEPYC 7601246810SE +/- 0.02, N = 36.37MIN: 5 / MAX: 6.49

OpenBenchmarking.orgFPS, More Is BetterOSPray 1.8.5Demo: Magnetic Reconnection - Renderer: Path TracerEPYC 76014080120160200200MIN: 111.11 / MAX: 250

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer - Model: CrownEPYC 7601510152025SE +/- 0.04, N = 320.95MIN: 20.62 / MAX: 21.36

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer ISPC - Model: CrownEPYC 7601510152025SE +/- 0.04, N = 319.54MIN: 19.21 / MAX: 19.93

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer - Model: Asian DragonEPYC 7601612182430SE +/- 0.03, N = 324.13MIN: 23.93 / MAX: 24.48

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer - Model: Asian Dragon ObjEPYC 7601510152025SE +/- 0.01, N = 321.59MIN: 21.41 / MAX: 21.92

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer ISPC - Model: Asian DragonEPYC 7601612182430SE +/- 0.14, N = 323.11MIN: 22.56 / MAX: 23.56

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 3.6.1Binary: Pathtracer ISPC - Model: Asian Dragon ObjEPYC 7601510152025SE +/- 0.05, N = 320.13MIN: 19.92 / MAX: 20.47

SVT-AV1

This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-AV1 CPU-based multi-threaded video encoder for the AV1 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.7Encoder Mode: Enc Mode 0 - Input: 1080pEPYC 76010.00950.0190.02850.0380.0475SE +/- 0.000, N = 60.0421. (CXX) g++ options: -fPIE -fPIC -pie

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.7Encoder Mode: Enc Mode 4 - Input: 1080pEPYC 76010.921.842.763.684.6SE +/- 0.037, N = 34.0891. (CXX) g++ options: -fPIE -fPIC -pie

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 0.7Encoder Mode: Enc Mode 8 - Input: 1080pEPYC 7601918273645SE +/- 0.19, N = 340.191. (CXX) g++ options: -fPIE -fPIC -pie

Intel Open Image Denoise

Open Image Denoise is a denoising library for ray-tracing and part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 1.0.0Scene: MemorialEPYC 7601246810SE +/- 0.03, N = 37.15

72 Results Shown

FFTW:
  Stock - 1D FFT Size 32
  Stock - 1D FFT Size 64
  Stock - 2D FFT Size 32
  Stock - 2D FFT Size 64
  Stock - 1D FFT Size 128
  Stock - 1D FFT Size 256
  Stock - 1D FFT Size 512
  Stock - 2D FFT Size 128
  Stock - 2D FFT Size 256
  Stock - 2D FFT Size 512
  Stock - 1D FFT Size 1024
  Stock - 1D FFT Size 2048
  Stock - 1D FFT Size 4096
  Stock - 2D FFT Size 1024
  Stock - 2D FFT Size 2048
  Stock - 2D FFT Size 4096
  Float + SSE - 1D FFT Size 32
  Float + SSE - 1D FFT Size 64
  Float + SSE - 2D FFT Size 32
  Float + SSE - 2D FFT Size 64
  Float + SSE - 1D FFT Size 128
  Float + SSE - 1D FFT Size 256
  Float + SSE - 1D FFT Size 512
  Float + SSE - 2D FFT Size 128
  Float + SSE - 2D FFT Size 256
  Float + SSE - 2D FFT Size 512
  Float + SSE - 1D FFT Size 1024
  Float + SSE - 1D FFT Size 2048
  Float + SSE - 1D FFT Size 4096
  Float + SSE - 2D FFT Size 1024
  Float + SSE - 2D FFT Size 2048
  Float + SSE - 2D FFT Size 4096
MKL-DNN DNNL:
  IP Batch 1D - f32
  IP Batch All - f32
  IP Batch 1D - u8s8f32
  IP Batch All - u8s8f32
  Convolution Batch conv_3d - f32
  Convolution Batch conv_all - f32
  Convolution Batch conv_3d - u8s8f32
  Deconvolution Batch deconv_1d - f32
  Deconvolution Batch deconv_3d - f32
  Convolution Batch conv_alexnet - f32
  Convolution Batch conv_all - u8s8f32
  Deconvolution Batch deconv_all - f32
  Deconvolution Batch deconv_1d - u8s8f32
  Deconvolution Batch deconv_3d - u8s8f32
  Recurrent Neural Network Training - f32
  Convolution Batch conv_alexnet - u8s8f32
  Convolution Batch conv_googlenet_v3 - f32
  Convolution Batch conv_googlenet_v3 - u8s8f32
dav1d:
  Chimera 1080p
  Summer Nature 4K
  Summer Nature 1080p
  Chimera 1080p 10-bit
OSPray:
  San Miguel - SciVis
  XFrog Forest - SciVis
  San Miguel - Path Tracer
  NASA Streamlines - SciVis
  XFrog Forest - Path Tracer
  Magnetic Reconnection - SciVis
  NASA Streamlines - Path Tracer
  Magnetic Reconnection - Path Tracer
Embree:
  Pathtracer - Crown
  Pathtracer ISPC - Crown
  Pathtracer - Asian Dragon
  Pathtracer - Asian Dragon Obj
  Pathtracer ISPC - Asian Dragon
  Pathtracer ISPC - Asian Dragon Obj
SVT-AV1:
  Enc Mode 0 - 1080p
  Enc Mode 4 - 1080p
  Enc Mode 8 - 1080p
Intel Open Image Denoise