general

2 x AMD EPYC 9274F 24-Core testing with a ASUS ESC8000A-E12 K14PG-D24 (1201 BIOS) and ASUS NVIDIA H100 NVL 94GB on Ubuntu 22.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2501312-NE-GENERAL1651
Jump To Table - Results

Statistics

Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core
January 29
  2 Days, 3 Hours, 48 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


generalOpenBenchmarking.orgPhoronix Test Suite2 x AMD EPYC 9274F 24-Core @ 4.05GHz (48 Cores / 96 Threads)ASUS ESC8000A-E12 K14PG-D24 (1201 BIOS)AMD Device 14a41136GB240GB MR9540-8iASUS NVIDIA H100 NVL 94GB2 x Broadcom BCM57414 NetXtreme-E 10Gb/25GbUbuntu 22.046.8.0-51-generic (x86_64)X ServerNVIDIAOpenCL 3.0 CUDA 12.7.331.3.289GCC 11.4.0 + CUDA 12.6ext41920x1200ProcessorMotherboardChipsetMemoryDiskGraphicsNetworkOSKernelDisplay ServerDisplay DriverOpenCLVulkanCompilerFile-SystemScreen ResolutionGeneral BenchmarksSystem Logs- Transparent Huge Pages: madvise- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-XeT9lY/gcc-11-11.4.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-XeT9lY/gcc-11-11.4.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - MQ-DEADLINE / relatime,rw,stripe=16 / Block Size: 4096- Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0xa101148 - Python 3.10.12- gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + reg_file_data_sampling: Not affected + retbleed: Not affected + spec_rstack_overflow: Mitigation of Safe RET + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced / Automatic IBRS; IBPB: conditional; STIBP: always-on; RSB filling; PBRSB-eIBRS: Not affected; BHI: Not affected + srbds: Not affected + tsx_async_abort: Not affected

generalllamafile: wizardcoder-python-34b-v1.0.Q6_K - Prompt Processing 2048llamafile: wizardcoder-python-34b-v1.0.Q6_K - Prompt Processing 1024llamafile: wizardcoder-python-34b-v1.0.Q6_K - Prompt Processing 512llamafile: wizardcoder-python-34b-v1.0.Q6_K - Prompt Processing 256llamafile: mistral-7b-instruct-v0.2.Q5_K_M - Prompt Processing 2048llamafile: mistral-7b-instruct-v0.2.Q5_K_M - Prompt Processing 1024llamafile: mistral-7b-instruct-v0.2.Q5_K_M - Prompt Processing 512llamafile: mistral-7b-instruct-v0.2.Q5_K_M - Prompt Processing 256llamafile: wizardcoder-python-34b-v1.0.Q6_K - Text Generation 128llamafile: TinyLlama-1.1B-Chat-v1.0.BF16 - Prompt Processing 2048llamafile: TinyLlama-1.1B-Chat-v1.0.BF16 - Prompt Processing 1024llamafile: wizardcoder-python-34b-v1.0.Q6_K - Text Generation 16llamafile: mistral-7b-instruct-v0.2.Q5_K_M - Text Generation 128llamafile: TinyLlama-1.1B-Chat-v1.0.BF16 - Prompt Processing 512llamafile: TinyLlama-1.1B-Chat-v1.0.BF16 - Prompt Processing 256llamafile: mistral-7b-instruct-v0.2.Q5_K_M - Text Generation 16llamafile: TinyLlama-1.1B-Chat-v1.0.BF16 - Text Generation 128llamafile: Llama-3.2-3B-Instruct.Q6_K - Prompt Processing 2048llamafile: Llama-3.2-3B-Instruct.Q6_K - Prompt Processing 1024llamafile: TinyLlama-1.1B-Chat-v1.0.BF16 - Text Generation 16llamafile: Llama-3.2-3B-Instruct.Q6_K - Prompt Processing 512llamafile: Llama-3.2-3B-Instruct.Q6_K - Prompt Processing 256llamafile: Llama-3.2-3B-Instruct.Q6_K - Text Generation 128llamafile: Llama-3.2-3B-Instruct.Q6_K - Text Generation 16llama-cpp: NVIDIA CUDA - granite-3.0-3b-a800m-instruct-Q8_0 - Prompt Processing 2048llama-cpp: NVIDIA CUDA - granite-3.0-3b-a800m-instruct-Q8_0 - Prompt Processing 1024llama-cpp: NVIDIA CUDA - granite-3.0-3b-a800m-instruct-Q8_0 - Prompt Processing 512llama-cpp: NVIDIA CUDA - granite-3.0-3b-a800m-instruct-Q8_0 - Text Generation 128llama-cpp: CPU BLAS - granite-3.0-3b-a800m-instruct-Q8_0 - Prompt Processing 2048llama-cpp: CPU BLAS - granite-3.0-3b-a800m-instruct-Q8_0 - Prompt Processing 1024llama-cpp: NVIDIA CUDA - Mistral-7B-Instruct-v0.3-Q8_0 - Prompt Processing 2048llama-cpp: NVIDIA CUDA - Mistral-7B-Instruct-v0.3-Q8_0 - Prompt Processing 1024llama-cpp: NVIDIA CUDA - Mistral-7B-Instruct-v0.3-Q8_0 - Prompt Processing 512llama-cpp: CPU BLAS - granite-3.0-3b-a800m-instruct-Q8_0 - Text Generation 128llama-cpp: NVIDIA CUDA - Mistral-7B-Instruct-v0.3-Q8_0 - Text Generation 128llama-cpp: CPU BLAS - Mistral-7B-Instruct-v0.3-Q8_0 - Prompt Processing 2048llama-cpp: CPU BLAS - Mistral-7B-Instruct-v0.3-Q8_0 - Prompt Processing 1024llama-cpp: CPU BLAS - Mistral-7B-Instruct-v0.3-Q8_0 - Prompt Processing 512llama-cpp: NVIDIA CUDA - Llama-3.1-Tulu-3-8B-Q8_0 - Prompt Processing 2048llama-cpp: NVIDIA CUDA - Llama-3.1-Tulu-3-8B-Q8_0 - Prompt Processing 1024llama-cpp: NVIDIA CUDA - Llama-3.1-Tulu-3-8B-Q8_0 - Prompt Processing 512llama-cpp: CPU BLAS - Mistral-7B-Instruct-v0.3-Q8_0 - Text Generation 128llama-cpp: NVIDIA CUDA - Llama-3.1-Tulu-3-8B-Q8_0 - Text Generation 128llama-cpp: CPU BLAS - Llama-3.1-Tulu-3-8B-Q8_0 - Prompt Processing 2048llama-cpp: CPU BLAS - Llama-3.1-Tulu-3-8B-Q8_0 - Prompt Processing 1024llama-cpp: CPU BLAS - Llama-3.1-Tulu-3-8B-Q8_0 - Prompt Processing 512llama-cpp: CPU BLAS - Llama-3.1-Tulu-3-8B-Q8_0 - Text Generation 128ai-benchmark: Device AI Scoreai-benchmark: Device Training Scoreai-benchmark: Device Inference Scorepybench: Total For Average Test Timesncnn: Vulkan GPU - vision_transformerncnn: Vulkan GPU - regnety_400mncnn: Vulkan GPU - squeezenet_ssdncnn: Vulkan GPU - yolov4-tinyncnn: Vulkan GPUv2-yolov3v2-yolov3 - mobilenetv2-yolov3ncnn: Vulkan GPU - resnet50ncnn: Vulkan GPU - alexnetncnn: Vulkan GPU - resnet18ncnn: Vulkan GPU - vgg16ncnn: Vulkan GPU - googlenetncnn: Vulkan GPU - blazefacencnn: Vulkan GPU - efficientnet-b0ncnn: Vulkan GPU - mnasnetncnn: Vulkan GPU - shufflenet-v2ncnn: Vulkan GPU-v3-v3 - mobilenet-v3ncnn: Vulkan GPU-v2-v2 - mobilenet-v2ncnn: Vulkan GPU - mobilenetncnn: CPU - vision_transformerncnn: CPU - regnety_400mncnn: CPU - squeezenet_ssdncnn: CPU - yolov4-tinyncnn: CPU - resnet50ncnn: CPU - alexnetncnn: CPU - resnet18ncnn: CPU - vgg16ncnn: CPU - googlenetncnn: CPU - blazefacencnn: CPU - efficientnet-b0ncnn: CPU - mnasnetncnn: CPU - shufflenet-v2ncnn: CPU-v2-v2 - mobilenet-v2spacy: en_core_web_lgstress-ng: Hyperbolic Trigonometric Mathstress-ng: POSIX Regular Expressionsstress-ng: System V Message Passingstress-ng: Glibc Qsort Data Sortingstress-ng: Glibc C String Functionsstress-ng: Integer Bit Operationsstress-ng: Bessel Math Operationsstress-ng: Vector Floating Pointstress-ng: Bitonic Integer Sortstress-ng: Trigonometric Mathstress-ng: Fused Multiply-Addstress-ng: Radix String Sortstress-ng: Fractal Generatorstress-ng: Context Switchingstress-ng: Wide Vector Mathstress-ng: Logarithmic Mathstress-ng: Jpeg Compressionstress-ng: Exponential Mathstress-ng: Socket Activitystress-ng: Mixed Schedulerstress-ng: Vector Shufflestress-ng: Memory Copyingstress-ng: Matrix 3D Mathstress-ng: Floating Pointstress-ng: x86_64 RdRandstress-ng: Function Callstress-ng: Integer Mathstress-ng: AVX-512 VNNIstress-ng: Vector Mathstress-ng: Matrix Mathstress-ng: Semaphoresstress-ng: Power Mathstress-ng: CPU Stressstress-ng: CPU Cachestress-ng: SENDFILEstress-ng: AVL Treestress-ng: Pthreadstress-ng: Forkingstress-ng: Cloningstress-ng: Mallocstress-ng: Atomicstress-ng: Mutexstress-ng: MEMFDstress-ng: Futexstress-ng: Zlibstress-ng: Pollstress-ng: Pipestress-ng: NUMAstress-ng: MMAPstress-ng: Hashpytorch: NVIDIA CUDA GPU - 512 - Efficientnet_v2_lpytorch: NVIDIA CUDA GPU - 256 - Efficientnet_v2_lpytorch: NVIDIA CUDA GPU - 64 - Efficientnet_v2_lpytorch: NVIDIA CUDA GPU - 32 - Efficientnet_v2_lpytorch: NVIDIA CUDA GPU - 16 - Efficientnet_v2_lpytorch: NVIDIA CUDA GPU - 1 - Efficientnet_v2_lpytorch: NVIDIA CUDA GPU - 512 - ResNet-152pytorch: NVIDIA CUDA GPU - 256 - ResNet-152pytorch: NVIDIA CUDA GPU - 64 - ResNet-152pytorch: NVIDIA CUDA GPU - 512 - ResNet-50pytorch: NVIDIA CUDA GPU - 32 - ResNet-152pytorch: NVIDIA CUDA GPU - 256 - ResNet-50pytorch: NVIDIA CUDA GPU - 16 - ResNet-152pytorch: NVIDIA CUDA GPU - 64 - ResNet-50pytorch: NVIDIA CUDA GPU - 32 - ResNet-50pytorch: NVIDIA CUDA GPU - 16 - ResNet-50pytorch: NVIDIA CUDA GPU - 1 - ResNet-152pytorch: NVIDIA CUDA GPU - 1 - ResNet-50pytorch: CPU - 512 - Efficientnet_v2_lpytorch: CPU - 256 - Efficientnet_v2_lpytorch: CPU - 64 - Efficientnet_v2_lpytorch: CPU - 32 - Efficientnet_v2_lpytorch: CPU - 16 - Efficientnet_v2_lpytorch: CPU - 1 - Efficientnet_v2_lpytorch: CPU - 512 - ResNet-152pytorch: CPU - 256 - ResNet-152pytorch: CPU - 64 - ResNet-152pytorch: CPU - 512 - ResNet-50pytorch: CPU - 32 - ResNet-152pytorch: CPU - 256 - ResNet-50pytorch: CPU - 16 - ResNet-152pytorch: CPU - 64 - ResNet-50pytorch: CPU - 32 - ResNet-50pytorch: CPU - 16 - ResNet-50pytorch: CPU - 1 - ResNet-152pytorch: CPU - 1 - ResNet-50tensorflow-lite: Mobilenet Quanttensorflow-lite: Mobilenet Floattensorflow-lite: NASNet Mobiletensorflow-lite: Inception V4tensorflow-lite: SqueezeNetlitert: Quantized COCO SSD MobileNet v1litert: Inception ResNet V2litert: Mobilenet Floatlitert: NASNet Mobilelitert: Inception V4litert: SqueezeNetlitert: DeepLab V3intel-mpi: IMB-MPI1 Sendrecvintel-mpi: IMB-MPI1 Sendrecvintel-mpi: IMB-MPI1 Exchangeintel-mpi: IMB-MPI1 Exchangeintel-mpi: IMB-P2P PingPongrbenchmark: cython-bench: N-Queensnumpy: onednn: Recurrent Neural Network Inference - CPUonednn: Recurrent Neural Network Training - CPUonednn: Deconvolution Batch shapes_3d - CPUonednn: Deconvolution Batch shapes_1d - CPUonednn: Convolution Batch Shapes Auto - CPUonednn: IP Shapes 3D - CPUonednn: IP Shapes 1D - CPUbuild-llvm: Unix Makefilesbuild-llvm: Ninjabuild-linux-kernel: allmodconfigbuild-linux-kernel: defconfigbuild-gcc: Time To Compileepoch: Conemrbayes: Primate Phylogeny Analysisnpb: SP.Cnpb: SP.Bnpb: MG.Cnpb: LU.Cnpb: IS.Dnpb: FT.Cnpb: EP.Dnpb: EP.Cnpb: CG.Cnpb: BT.Chpl: glibc-bench: pthread_onceglibc-bench: sincosglibc-bench: ffsllglibc-bench: atanhglibc-bench: asinhglibc-bench: tanhglibc-bench: sqrtglibc-bench: sinhglibc-bench: modfglibc-bench: log2glibc-bench: singlibc-bench: powglibc-bench: ffsglibc-bench: expglibc-bench: coscompilebench: Read Compiled Treecompilebench: Initial Createcompilebench: Compilellama-cpp: CPU BLAS - granite-3.0-3b-a800m-instruct-Q8_0 - Prompt Processing 512ncnn: Vulkan GPU - FastestDetncnn: CPU - FastestDetncnn: CPUv2-yolov3v2-yolov3 - mobilenetv2-yolov3ncnn: CPU-v3-v3 - mobilenet-v3ncnn: CPU - mobilenetspacy: en_core_web_trfstress-ng: IO_uringstress-ng: Cryptotensorflow-lite: Inception ResNet V2litert: Mobilenet Quantintel-mpi: IMB-MPI1 PingPongUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core122886144307215363276816384819240965.0032768163844.9224.128192409624.0456.82327681638457.788192409642.1641.913121.143150.783144.4284.98211.93212.9218308.6113052.608390.8363.08151.8344.0042.8843.5418283.7513019.738352.5718.58147.6345.5743.8543.8717.5665763380319673856.6546.6723.7231.3023.1522.636.5812.4641.4122.256.3516.9911.6916.5613.3311.7723.1557.2145.7323.9831.7522.826.3212.6942.1222.656.3916.7111.5216.7611.8515338293490.26444669.6014211326.191590.9156864254.149623549.8336898.14182971.03672.51151938.4247688547.662158.48393.6715485954.922734553.44374205.2266045.55198051.520499.6638187.3942786.9013628.1416067.5320022.5120427852.8245580.454586743.656540887.21407859.79310400.3161709090.58119629.76145536.61771976.41916359.111194.06135182.4062039.408039.79106614259.32257.0214723637.913377.152352726.226888.225524301.7121174901.61400.8012391.7012298718.4346.7247.3447.8046.7747.3251.52102.98102.81102.21283.02102.43282.30102.11279.19282.75279.17101.94287.466.096.076.186.055.998.9211.6111.4511.5130.0512.2929.0111.2729.0629.0329.7713.9734.934098.483350.7269069.737181.95088.145503.9145507.63345.8575527.338062.45245.278793.8779.473393.59108.056073.65287445890.144617.977565.16398.494669.6831.103288.665020.6733690.3108890.827742198.897135.289258.63730.299913.620384.1981.583108596.64146367.34119325.63199460.274064.08102149.257273.966692.9550419.73176096.16631.305.6554738.98055.6705528.175722.583026.97268.2457422.95506.6195710.421563.122735.56075.6703215.136170.81582657.42399.791613.38193.8117.2917.7923.3114.3623.312751174744.18402622178.2868557.93255.763739.53OpenBenchmarking.org

Llamafile

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: wizardcoder-python-34b-v1.0.Q6_K - Test: Prompt Processing 2048Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3K6K9K12K15KSE +/- 0.00, N = 312288

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: wizardcoder-python-34b-v1.0.Q6_K - Test: Prompt Processing 1024Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core13002600390052006500SE +/- 0.00, N = 36144

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: wizardcoder-python-34b-v1.0.Q6_K - Test: Prompt Processing 512Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core7001400210028003500SE +/- 0.00, N = 33072

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: wizardcoder-python-34b-v1.0.Q6_K - Test: Prompt Processing 256Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core30060090012001500SE +/- 0.00, N = 31536

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: mistral-7b-instruct-v0.2.Q5_K_M - Test: Prompt Processing 2048Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core7K14K21K28K35KSE +/- 0.00, N = 332768

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: mistral-7b-instruct-v0.2.Q5_K_M - Test: Prompt Processing 1024Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core4K8K12K16K20KSE +/- 0.00, N = 316384

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: mistral-7b-instruct-v0.2.Q5_K_M - Test: Prompt Processing 512Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core2K4K6K8K10KSE +/- 0.00, N = 38192

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: mistral-7b-instruct-v0.2.Q5_K_M - Test: Prompt Processing 256Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core9001800270036004500SE +/- 0.00, N = 34096

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: wizardcoder-python-34b-v1.0.Q6_K - Test: Text Generation 128Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1.1252.253.3754.55.625SE +/- 0.02, N = 35.00

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: TinyLlama-1.1B-Chat-v1.0.BF16 - Test: Prompt Processing 2048Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core7K14K21K28K35KSE +/- 0.00, N = 332768

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: TinyLlama-1.1B-Chat-v1.0.BF16 - Test: Prompt Processing 1024Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core4K8K12K16K20KSE +/- 0.00, N = 316384

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: wizardcoder-python-34b-v1.0.Q6_K - Test: Text Generation 16Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1.1072.2143.3214.4285.535SE +/- 0.03, N = 34.92

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: mistral-7b-instruct-v0.2.Q5_K_M - Test: Text Generation 128Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core612182430SE +/- 0.04, N = 324.12

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: TinyLlama-1.1B-Chat-v1.0.BF16 - Test: Prompt Processing 512Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core2K4K6K8K10KSE +/- 0.00, N = 38192

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: TinyLlama-1.1B-Chat-v1.0.BF16 - Test: Prompt Processing 256Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core9001800270036004500SE +/- 0.00, N = 34096

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: mistral-7b-instruct-v0.2.Q5_K_M - Test: Text Generation 16Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core612182430SE +/- 0.09, N = 324.04

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: TinyLlama-1.1B-Chat-v1.0.BF16 - Test: Text Generation 128Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1326395265SE +/- 0.59, N = 356.82

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: Llama-3.2-3B-Instruct.Q6_K - Test: Prompt Processing 2048Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core7K14K21K28K35KSE +/- 0.00, N = 332768

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: Llama-3.2-3B-Instruct.Q6_K - Test: Prompt Processing 1024Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core4K8K12K16K20KSE +/- 0.00, N = 316384

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: TinyLlama-1.1B-Chat-v1.0.BF16 - Test: Text Generation 16Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1326395265SE +/- 0.61, N = 457.78

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: Llama-3.2-3B-Instruct.Q6_K - Test: Prompt Processing 512Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core2K4K6K8K10KSE +/- 0.00, N = 38192

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: Llama-3.2-3B-Instruct.Q6_K - Test: Prompt Processing 256Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core9001800270036004500SE +/- 0.00, N = 34096

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: Llama-3.2-3B-Instruct.Q6_K - Test: Text Generation 128Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1020304050SE +/- 0.30, N = 342.16

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.16Model: Llama-3.2-3B-Instruct.Q6_K - Test: Text Generation 16Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1020304050SE +/- 0.21, N = 341.91

Llama.cpp

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: NVIDIA CUDA - Model: granite-3.0-3b-a800m-instruct-Q8_0 - Test: Prompt Processing 2048Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core7001400210028003500SE +/- 0.10, N = 33121.141. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: NVIDIA CUDA - Model: granite-3.0-3b-a800m-instruct-Q8_0 - Test: Prompt Processing 1024Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core7001400210028003500SE +/- 0.63, N = 33150.781. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: NVIDIA CUDA - Model: granite-3.0-3b-a800m-instruct-Q8_0 - Test: Prompt Processing 512Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core7001400210028003500SE +/- 0.28, N = 33144.421. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: NVIDIA CUDA - Model: granite-3.0-3b-a800m-instruct-Q8_0 - Test: Text Generation 128Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core20406080100SE +/- 0.01, N = 384.981. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: CPU BLAS - Model: granite-3.0-3b-a800m-instruct-Q8_0 - Test: Prompt Processing 2048Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core50100150200250SE +/- 2.80, N = 3211.931. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: CPU BLAS - Model: granite-3.0-3b-a800m-instruct-Q8_0 - Test: Prompt Processing 1024Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core50100150200250SE +/- 2.68, N = 15212.921. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: NVIDIA CUDA - Model: Mistral-7B-Instruct-v0.3-Q8_0 - Test: Prompt Processing 2048Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core4K8K12K16K20KSE +/- 1.84, N = 318308.611. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: NVIDIA CUDA - Model: Mistral-7B-Instruct-v0.3-Q8_0 - Test: Prompt Processing 1024Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3K6K9K12K15KSE +/- 1.13, N = 313052.601. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: NVIDIA CUDA - Model: Mistral-7B-Instruct-v0.3-Q8_0 - Test: Prompt Processing 512Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core2K4K6K8K10KSE +/- 0.48, N = 38390.831. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: CPU BLAS - Model: granite-3.0-3b-a800m-instruct-Q8_0 - Test: Text Generation 128Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1428425670SE +/- 0.35, N = 363.081. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: NVIDIA CUDA - Model: Mistral-7B-Instruct-v0.3-Q8_0 - Test: Text Generation 128Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core306090120150SE +/- 0.10, N = 3151.831. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: CPU BLAS - Model: Mistral-7B-Instruct-v0.3-Q8_0 - Test: Prompt Processing 2048Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1020304050SE +/- 0.84, N = 944.001. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: CPU BLAS - Model: Mistral-7B-Instruct-v0.3-Q8_0 - Test: Prompt Processing 1024Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1020304050SE +/- 0.38, N = 1242.881. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: CPU BLAS - Model: Mistral-7B-Instruct-v0.3-Q8_0 - Test: Prompt Processing 512Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1020304050SE +/- 0.69, N = 1243.541. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: NVIDIA CUDA - Model: Llama-3.1-Tulu-3-8B-Q8_0 - Test: Prompt Processing 2048Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core4K8K12K16K20KSE +/- 0.46, N = 318283.751. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: NVIDIA CUDA - Model: Llama-3.1-Tulu-3-8B-Q8_0 - Test: Prompt Processing 1024Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3K6K9K12K15KSE +/- 5.36, N = 313019.731. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: NVIDIA CUDA - Model: Llama-3.1-Tulu-3-8B-Q8_0 - Test: Prompt Processing 512Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core2K4K6K8K10KSE +/- 0.80, N = 38352.571. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: CPU BLAS - Model: Mistral-7B-Instruct-v0.3-Q8_0 - Test: Text Generation 128Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core510152025SE +/- 0.03, N = 318.581. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: NVIDIA CUDA - Model: Llama-3.1-Tulu-3-8B-Q8_0 - Test: Text Generation 128Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core306090120150SE +/- 0.09, N = 3147.631. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: CPU BLAS - Model: Llama-3.1-Tulu-3-8B-Q8_0 - Test: Prompt Processing 2048Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1020304050SE +/- 0.99, N = 745.571. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: CPU BLAS - Model: Llama-3.1-Tulu-3-8B-Q8_0 - Test: Prompt Processing 1024Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1020304050SE +/- 0.75, N = 1243.851. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: CPU BLAS - Model: Llama-3.1-Tulu-3-8B-Q8_0 - Test: Prompt Processing 512Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1020304050SE +/- 0.40, N = 1543.871. (CXX) g++ options: -O3

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: CPU BLAS - Model: Llama-3.1-Tulu-3-8B-Q8_0 - Test: Text Generation 128Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core48121620SE +/- 0.02, N = 317.561. (CXX) g++ options: -O3

AI Benchmark Alpha

AI Benchmark Alpha is a Python library for evaluating artificial intelligence (AI) performance on diverse hardware platforms and relies upon the TensorFlow machine learning library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgScore, More Is BetterAI Benchmark Alpha 0.1.2Device AI ScoreUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core140028004200560070006576

OpenBenchmarking.orgScore, More Is BetterAI Benchmark Alpha 0.1.2Device Training ScoreUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core70014002100280035003380

OpenBenchmarking.orgScore, More Is BetterAI Benchmark Alpha 0.1.2Device Inference ScoreUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core70014002100280035003196

PyBench

This test profile reports the total time of the different average timed test results from PyBench. PyBench reports average test times for different functions such as BuiltinFunctionCalls and NestedForLoops, with this total result providing a rough estimate as to Python's average performance on a given system. This test profile runs PyBench each time for 20 rounds. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyBench 2018-02-16Total For Average Test TimesUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core160320480640800SE +/- 3.06, N = 3738

NCNN

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU - Model: vision_transformerUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1326395265SE +/- 0.48, N = 356.65MIN: 54.43 / MAX: 62.811. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU - Model: regnety_400mUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1122334455SE +/- 0.91, N = 346.67MIN: 45.2 / MAX: 88.321. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU - Model: squeezenet_ssdUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core612182430SE +/- 0.30, N = 323.72MIN: 23.06 / MAX: 29.241. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU - Model: yolov4-tinyUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core714212835SE +/- 0.69, N = 331.30MIN: 29.76 / MAX: 37.031. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPUv2-yolov3v2-yolov3 - Model: mobilenetv2-yolov3Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core612182430SE +/- 0.17, N = 323.15MIN: 22.82 / MAX: 28.931. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU - Model: resnet50Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core510152025SE +/- 0.47, N = 322.63MIN: 21.57 / MAX: 30.391. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU - Model: alexnetUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core246810SE +/- 0.16, N = 36.58MIN: 6.19 / MAX: 11.021. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU - Model: resnet18Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3691215SE +/- 0.18, N = 312.46MIN: 11.95 / MAX: 13.131. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU - Model: vgg16Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core918273645SE +/- 1.07, N = 341.41MIN: 39.19 / MAX: 48.361. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU - Model: googlenetUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core510152025SE +/- 0.24, N = 322.25MIN: 21.82 / MAX: 28.731. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU - Model: blazefaceUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core246810SE +/- 0.05, N = 36.35MIN: 6.16 / MAX: 11.261. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU - Model: efficientnet-b0Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core48121620SE +/- 0.15, N = 316.99MIN: 16.56 / MAX: 22.21. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU - Model: mnasnetUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3691215SE +/- 0.22, N = 311.69MIN: 11.09 / MAX: 19.421. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU - Model: shufflenet-v2Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core48121620SE +/- 0.09, N = 316.56MIN: 16.28 / MAX: 17.551. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU-v3-v3 - Model: mobilenet-v3Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3691215SE +/- 0.23, N = 313.33MIN: 12.65 / MAX: 18.551. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU-v2-v2 - Model: mobilenet-v2Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3691215SE +/- 0.40, N = 311.77MIN: 11.14 / MAX: 13.151. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU - Model: mobilenetUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core612182430SE +/- 0.17, N = 323.15MIN: 22.82 / MAX: 28.931. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU - Model: vision_transformerUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1326395265SE +/- 0.39, N = 1257.21MIN: 52 / MAX: 568.961. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU - Model: regnety_400mUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1020304050SE +/- 0.27, N = 1245.73MIN: -425.63 / MAX: 85.351. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU - Model: squeezenet_ssdUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core612182430SE +/- 0.32, N = 1223.98MIN: 21.56 / MAX: 65.281. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU - Model: yolov4-tinyUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core714212835SE +/- 0.47, N = 1231.75MIN: 28.18 / MAX: 38.621. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU - Model: resnet50Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core510152025SE +/- 0.25, N = 1222.82MIN: 21.27 / MAX: 32.641. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU - Model: alexnetUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core246810SE +/- 0.07, N = 126.32MIN: 5.67 / MAX: 19.421. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU - Model: resnet18Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3691215SE +/- 0.15, N = 1212.69MIN: 11.43 / MAX: 28.061. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU - Model: vgg16Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1020304050SE +/- 0.44, N = 1242.12MIN: 38.09 / MAX: 61.811. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU - Model: googlenetUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core510152025SE +/- 0.26, N = 1222.65MIN: 20.21 / MAX: 53.241. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU - Model: blazefaceUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core246810SE +/- 0.05, N = 126.39MIN: 5.9 / MAX: 7.261. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU - Model: efficientnet-b0Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core48121620SE +/- 0.14, N = 1216.71MIN: 14.94 / MAX: 30.141. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU - Model: mnasnetUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3691215SE +/- 0.07, N = 1211.52MIN: 10.69 / MAX: 16.831. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU - Model: shufflenet-v2Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core48121620SE +/- 0.05, N = 1216.76MIN: 16.18 / MAX: 40.161. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU-v2-v2 - Model: mobilenet-v2Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3691215SE +/- 0.07, N = 1211.85MIN: 11.19 / MAX: 16.921. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

spaCy

The spaCy library is an open-source solution for advanced neural language processing (NLP). The spaCy library leverages Python and is a leading neural language processing solution. This test profile times the spaCy CPU performance with various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgtokens/sec, More Is BetterspaCy 3.4.1Model: en_core_web_lgUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3K6K9K12K15KSE +/- 33.20, N = 315338

Stress-NG

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Hyperbolic Trigonometric MathUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core60K120K180K240K300KSE +/- 78.25, N = 3293490.261. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: POSIX Regular ExpressionsUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core100K200K300K400K500KSE +/- 185.23, N = 3444669.601. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: System V Message PassingUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3M6M9M12M15MSE +/- 27048.08, N = 314211326.191. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Glibc Qsort Data SortingUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core30060090012001500SE +/- 0.21, N = 31590.911. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Glibc C String FunctionsUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core12M24M36M48M60MSE +/- 270620.44, N = 356864254.141. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Integer Bit OperationsUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core2M4M6M8M10MSE +/- 1390.58, N = 39623549.831. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Bessel Math OperationsUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core8K16K24K32K40KSE +/- 2.68, N = 336898.141. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Vector Floating PointUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core40K80K120K160K200KSE +/- 489.30, N = 3182971.031. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Bitonic Integer SortUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core150300450600750SE +/- 0.35, N = 3672.511. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Trigonometric MathUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core30K60K90K120K150KSE +/- 47.69, N = 3151938.421. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Fused Multiply-AddUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core10M20M30M40M50MSE +/- 16313.26, N = 347688547.661. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Radix String SortUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core5001000150020002500SE +/- 14.11, N = 32158.481. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Fractal GeneratorUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core90180270360450SE +/- 0.18, N = 3393.671. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Context SwitchingUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3M6M9M12M15MSE +/- 27120.62, N = 315485954.921. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Wide Vector MathUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core600K1200K1800K2400K3000KSE +/- 706.91, N = 32734553.441. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Logarithmic MathUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core80K160K240K320K400KSE +/- 588.05, N = 3374205.221. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Jpeg CompressionUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core14K28K42K56K70KSE +/- 141.87, N = 366045.551. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Exponential MathUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core40K80K120K160K200KSE +/- 1213.20, N = 3198051.51. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Socket ActivityUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core4K8K12K16K20KSE +/- 5.10, N = 320499.661. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Mixed SchedulerUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core8K16K24K32K40KSE +/- 172.62, N = 338187.391. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Vector ShuffleUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core9K18K27K36K45KSE +/- 211.62, N = 342786.901. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Memory CopyingUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3K6K9K12K15KSE +/- 18.97, N = 313628.141. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Matrix 3D MathUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3K6K9K12K15KSE +/- 58.20, N = 316067.531. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Floating PointUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core4K8K12K16K20KSE +/- 76.48, N = 320022.511. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: x86_64 RdRandUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core4M8M12M16M20MSE +/- 68905.39, N = 320427852.821. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Function CallUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core10K20K30K40K50KSE +/- 71.73, N = 345580.451. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Integer MathUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1000K2000K3000K4000K5000KSE +/- 19349.76, N = 34586743.651. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: AVX-512 VNNIUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1.4M2.8M4.2M5.6M7MSE +/- 25442.23, N = 36540887.211. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Vector MathUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core90K180K270K360K450KSE +/- 2733.46, N = 3407859.791. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Matrix MathUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core70K140K210K280K350KSE +/- 26.15, N = 3310400.311. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: SemaphoresUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core13M26M39M52M65MSE +/- 563816.56, N = 361709090.581. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: Power MathUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core30K60K90K120K150KSE +/- 87.83, N = 3119629.761. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: CPU StressUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core30K60K90K120K150KSE +/- 271.68, N = 3145536.611. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: CPU CacheUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core170K340K510K680K850KSE +/- 2865.93, N = 3771976.411. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: SENDFILEUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core200K400K600K800K1000KSE +/- 380.91, N = 3916359.111. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: AVL TreeUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core30060090012001500SE +/- 1.00, N = 31194.061. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: PthreadUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core30K60K90K120K150KSE +/- 113.35, N = 3135182.401. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: ForkingUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core13K26K39K52K65KSE +/- 359.02, N = 362039.401. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: CloningUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core2K4K6K8K10KSE +/- 68.82, N = 38039.791. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: MallocUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core20M40M60M80M100MSE +/- 781395.83, N = 3106614259.321. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: AtomicUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core60120180240300SE +/- 0.39, N = 3257.021. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: MutexUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3M6M9M12M15MSE +/- 89617.88, N = 314723637.911. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: MEMFDUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core7001400210028003500SE +/- 3.40, N = 33377.151. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: FutexUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core500K1000K1500K2000K2500KSE +/- 20321.47, N = 32352726.221. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: ZlibUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core15003000450060007500SE +/- 2.93, N = 36888.221. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: PollUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1.2M2.4M3.6M4.8M6MSE +/- 2718.46, N = 35524301.711. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: PipeUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core5M10M15M20M25MSE +/- 278876.82, N = 1221174901.611. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: NUMAUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core90180270360450SE +/- 2.65, N = 3400.801. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: MMAPUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3K6K9K12K15KSE +/- 89.55, N = 312391.701. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: HashUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3M6M9M12M15MSE +/- 46080.65, N = 312298718.431. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

PyTorch

This is a benchmark of PyTorch making use of pytorch-benchmark [https://github.com/LukasHedegaard/pytorch-benchmark]. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 512 - Model: Efficientnet_v2_lUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1122334455SE +/- 0.07, N = 346.72MIN: 39.24 / MAX: 48.63

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 256 - Model: Efficientnet_v2_lUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1122334455SE +/- 0.50, N = 347.34MIN: 39.26 / MAX: 50.11

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 64 - Model: Efficientnet_v2_lUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1122334455SE +/- 0.44, N = 347.80MIN: 2.23 / MAX: 49.81

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 32 - Model: Efficientnet_v2_lUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1122334455SE +/- 0.61, N = 346.77MIN: 36.59 / MAX: 48.4

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 16 - Model: Efficientnet_v2_lUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1122334455SE +/- 0.24, N = 347.32MIN: 36.64 / MAX: 48.53

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 1 - Model: Efficientnet_v2_lUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1224364860SE +/- 0.62, N = 451.52MIN: 36.53 / MAX: 55.15

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 512 - Model: ResNet-152Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core20406080100SE +/- 0.47, N = 3102.98MIN: 73.53 / MAX: 107.1

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 256 - Model: ResNet-152Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core20406080100SE +/- 0.62, N = 3102.81MIN: -2.4 / MAX: 105.21

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 64 - Model: ResNet-152Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core20406080100SE +/- 0.84, N = 3102.21MIN: 73.94 / MAX: 104.89

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 512 - Model: ResNet-50Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core60120180240300SE +/- 3.04, N = 3283.02MIN: 168.78 / MAX: 291.41

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 32 - Model: ResNet-152Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core20406080100SE +/- 1.05, N = 3102.43MIN: 75.28 / MAX: 104.92

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 256 - Model: ResNet-50Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core60120180240300SE +/- 3.61, N = 3282.30MIN: 166.48 / MAX: 290.65

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 16 - Model: ResNet-152Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core20406080100SE +/- 0.30, N = 3102.11MIN: 74.46 / MAX: 105.04

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 64 - Model: ResNet-50Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core60120180240300SE +/- 1.76, N = 3279.19MIN: 167.62 / MAX: 285.62

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 32 - Model: ResNet-50Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core60120180240300SE +/- 2.42, N = 3282.75MIN: 166.76 / MAX: 292.39

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 16 - Model: ResNet-50Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core60120180240300SE +/- 2.06, N = 3279.17MIN: 168.32 / MAX: 287.01

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 1 - Model: ResNet-152Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core20406080100SE +/- 1.00, N = 3101.94MIN: 72.8 / MAX: 104.88

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: NVIDIA CUDA GPU - Batch Size: 1 - Model: ResNet-50Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core60120180240300SE +/- 3.16, N = 3287.46MIN: 131.87 / MAX: 296.46

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 512 - Model: Efficientnet_v2_lUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core246810SE +/- 0.06, N = 36.09MIN: 1.78 / MAX: 6.47

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 256 - Model: Efficientnet_v2_lUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core246810SE +/- 0.06, N = 66.07MIN: 3.72 / MAX: 6.69

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 64 - Model: Efficientnet_v2_lUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core246810SE +/- 0.06, N = 36.18MIN: 4.86 / MAX: 6.54

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 32 - Model: Efficientnet_v2_lUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core246810SE +/- 0.04, N = 36.05MIN: 4.81 / MAX: 6.36

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 16 - Model: Efficientnet_v2_lUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1.34782.69564.04345.39126.739SE +/- 0.05, N = 35.99MIN: 4.56 / MAX: 6.36

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 1 - Model: Efficientnet_v2_lUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core246810SE +/- 0.13, N = 38.92MIN: 5.44 / MAX: 9.23

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 512 - Model: ResNet-152Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3691215SE +/- 0.09, N = 1211.61MIN: 6.77 / MAX: 12.65

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 256 - Model: ResNet-152Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3691215SE +/- 0.03, N = 311.45MIN: 10.75 / MAX: 12.11

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 64 - Model: ResNet-152Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3691215SE +/- 0.12, N = 311.51MIN: 7.49 / MAX: 12.29

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 512 - Model: ResNet-50Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core714212835SE +/- 0.44, N = 1230.05MIN: 17.64 / MAX: 35.18

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 32 - Model: ResNet-152Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3691215SE +/- 0.13, N = 312.29MIN: 7.38 / MAX: 12.68

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 256 - Model: ResNet-50Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core714212835SE +/- 0.32, N = 529.01MIN: 16.97 / MAX: 32.29

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 16 - Model: ResNet-152Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3691215SE +/- 0.11, N = 1211.27MIN: 6.86 / MAX: 12.45

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 64 - Model: ResNet-50Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core714212835SE +/- 0.18, N = 329.06MIN: 16.5 / MAX: 32.6

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 32 - Model: ResNet-50Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core714212835SE +/- 0.21, N = 329.03MIN: 19.55 / MAX: 32.27

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 16 - Model: ResNet-50Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core714212835SE +/- 0.40, N = 1529.77MIN: 13.43 / MAX: 34.1

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 1 - Model: ResNet-152Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core48121620SE +/- 0.11, N = 313.97MIN: 8.7 / MAX: 15.19

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 1 - Model: ResNet-50Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core816243240SE +/- 0.39, N = 1534.93MIN: 13.13 / MAX: 41.07

TensorFlow Lite

This is a benchmark of the TensorFlow Lite implementation focused on TensorFlow machine learning for mobile, IoT, edge, and other cases. The current Linux support is limited to running on CPUs. This test profile is measuring the average inference time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: Mobilenet QuantUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core9001800270036004500SE +/- 12.05, N = 34098.48

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: Mobilenet FloatUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core7001400210028003500SE +/- 26.31, N = 153350.72

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: NASNet MobileUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core15K30K45K60K75KSE +/- 562.50, N = 369069.7

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: Inception V4Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core8K16K24K32K40KSE +/- 358.95, N = 337181.9

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: SqueezeNetUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core11002200330044005500SE +/- 29.41, N = 35088.14

LiteRT

OpenBenchmarking.orgMicroseconds, Fewer Is BetterLiteRT 2024-10-15Model: Quantized COCO SSD MobileNet v1Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core12002400360048006000SE +/- 71.29, N = 135503.91

OpenBenchmarking.orgMicroseconds, Fewer Is BetterLiteRT 2024-10-15Model: Inception ResNet V2Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core10K20K30K40K50KSE +/- 312.59, N = 1545507.6

OpenBenchmarking.orgMicroseconds, Fewer Is BetterLiteRT 2024-10-15Model: Mobilenet FloatUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core7001400210028003500SE +/- 39.32, N = 33345.85

OpenBenchmarking.orgMicroseconds, Fewer Is BetterLiteRT 2024-10-15Model: NASNet MobileUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core16K32K48K64K80KSE +/- 699.04, N = 775527.3

OpenBenchmarking.orgMicroseconds, Fewer Is BetterLiteRT 2024-10-15Model: Inception V4Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core8K16K24K32K40KSE +/- 455.18, N = 338062.4

OpenBenchmarking.orgMicroseconds, Fewer Is BetterLiteRT 2024-10-15Model: SqueezeNetUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core11002200330044005500SE +/- 33.72, N = 35245.27

OpenBenchmarking.orgMicroseconds, Fewer Is BetterLiteRT 2024-10-15Model: DeepLab V3Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core2K4K6K8K10KSE +/- 57.34, N = 158793.87

Intel MPI Benchmarks

Intel MPI Benchmarks for stressing MPI implementations. At this point the test profile aggregates results for some common MPI functionality. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgAverage usec, Fewer Is BetterIntel MPI Benchmarks 2019.3Test: IMB-MPI1 SendrecvUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core20406080100SE +/- 0.66, N = 379.47MIN: 0.64 / MAX: 1174.191. (CXX) g++ options: -O0 -pedantic -fopenmp -lmpi_cxx -lmpi

OpenBenchmarking.orgAverage Mbytes/sec, More Is BetterIntel MPI Benchmarks 2019.3Test: IMB-MPI1 SendrecvUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core7001400210028003500SE +/- 22.79, N = 33393.59MAX: 13584.141. (CXX) g++ options: -O0 -pedantic -fopenmp -lmpi_cxx -lmpi

OpenBenchmarking.orgAverage usec, Fewer Is BetterIntel MPI Benchmarks 2019.3Test: IMB-MPI1 ExchangeUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core20406080100SE +/- 3.64, N = 3108.05MIN: 1.13 / MAX: 1949.291. (CXX) g++ options: -O0 -pedantic -fopenmp -lmpi_cxx -lmpi

OpenBenchmarking.orgAverage Mbytes/sec, More Is BetterIntel MPI Benchmarks 2019.3Test: IMB-MPI1 ExchangeUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core13002600390052006500SE +/- 18.92, N = 36073.65MAX: 26527.081. (CXX) g++ options: -O0 -pedantic -fopenmp -lmpi_cxx -lmpi

OpenBenchmarking.orgAverage Msg/sec, More Is BetterIntel MPI Benchmarks 2019.3Test: IMB-P2P PingPongUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core6M12M18M24M30MSE +/- 68029.72, N = 328744589MIN: 15552 / MAX: 713885661. (CXX) g++ options: -O0 -pedantic -fopenmp -lmpi_cxx -lmpi

R Benchmark

This test is a quick-running survey of general R performance Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterR BenchmarkUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core0.03250.0650.09750.130.1625SE +/- 0.0014, N = 30.14461. R scripting front-end version 4.1.2 (2021-11-01)

Cython Benchmark

Cython provides a superset of Python that is geared to deliver C-like levels of performance. This test profile makes use of Cython's bundled benchmark tests and runs an N-Queens sample test as a simple benchmark to the system's Cython performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterCython Benchmark 0.29.21Test: N-QueensUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core48121620SE +/- 0.14, N = 317.98

Numpy Benchmark

This is a test to obtain the general Numpy performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgScore, More Is BetterNumpy BenchmarkUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core120240360480600SE +/- 0.45, N = 3565.16

oneDNN

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.6Harness: Recurrent Neural Network Inference - Engine: CPUUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core90180270360450SE +/- 0.88, N = 3398.49MIN: 390.221. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -fcf-protection=full -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.6Harness: Recurrent Neural Network Training - Engine: CPUUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core140280420560700SE +/- 1.20, N = 3669.68MIN: 653.951. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -fcf-protection=full -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.6Harness: Deconvolution Batch shapes_3d - Engine: CPUUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core0.24820.49640.74460.99281.241SE +/- 0.00100, N = 31.10328MIN: 1.051. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -fcf-protection=full -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.6Harness: Deconvolution Batch shapes_1d - Engine: CPUUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core246810SE +/- 0.03403, N = 38.66502MIN: 7.331. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -fcf-protection=full -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.6Harness: Convolution Batch Shapes Auto - Engine: CPUUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core0.15150.3030.45450.6060.7575SE +/- 0.001462, N = 30.673369MIN: 0.641. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -fcf-protection=full -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.6Harness: IP Shapes 3D - Engine: CPUUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core0.070.140.210.280.35SE +/- 0.002020, N = 30.310889MIN: 0.281. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -fcf-protection=full -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.6Harness: IP Shapes 1D - Engine: CPUUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core0.18620.37240.55860.74480.931SE +/- 0.005050, N = 30.827742MIN: 0.771. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -fcf-protection=full -pie -ldl

Timed LLVM Compilation

This test times how long it takes to compile/build the LLVM compiler stack. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed LLVM Compilation 16.0Build System: Unix MakefilesUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core4080120160200SE +/- 1.10, N = 3198.90

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed LLVM Compilation 16.0Build System: NinjaUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core306090120150SE +/- 1.19, N = 3135.29

Timed Linux Kernel Compilation

This test times how long it takes to build the Linux kernel in a default configuration (defconfig) for the architecture being tested or alternatively an allmodconfig for building all possible kernel modules for the build. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 6.8Build: allmodconfigUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core60120180240300SE +/- 0.84, N = 3258.64

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 6.8Build: defconfigUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core714212835SE +/- 0.22, N = 1430.30

Timed GCC Compilation

This test times how long it takes to build the GNU Compiler Collection (GCC) open-source compiler. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed GCC Compilation 13.2Time To CompileUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core2004006008001000SE +/- 0.79, N = 3913.62

Epoch

OpenBenchmarking.orgSeconds, Fewer Is BetterEpoch 4.19.4Epoch3D Deck: ConeUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core80160240320400SE +/- 4.25, N = 3384.191. (F9X) gfortran options: -O3 -std=f2003 -Jobj -lsdf -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz

Timed MrBayes Analysis

This test performs a bayesian analysis of a set of primate genome sequences in order to estimate their phylogeny. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed MrBayes Analysis 3.2.7Primate Phylogeny AnalysisUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core20406080100SE +/- 0.77, N = 381.581. (CC) gcc options: -mmmx -msse -msse2 -msse3 -mssse3 -msse4.1 -msse4.2 -msse4a -msha -maes -mavx -mfma -mavx2 -mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mavx512ifma -mavx512vbmi -mrdrnd -mbmi -mbmi2 -madx -mabm -O3 -std=c99 -pedantic -lm -lreadline

NAS Parallel Benchmarks

NPB, NAS Parallel Benchmarks, is a benchmark developed by NASA for high-end computer systems. This test profile currently uses the MPI version of NPB. This test profile offers selecting the different NPB tests/problems and varying problem sizes. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: SP.CUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core20K40K60K80K100KSE +/- 153.92, N = 3108596.641. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.2

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: SP.BUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core30K60K90K120K150KSE +/- 1283.80, N = 8146367.341. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.2

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: MG.CUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core30K60K90K120K150KSE +/- 947.46, N = 3119325.631. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.2

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: LU.CUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core40K80K120K160K200KSE +/- 202.80, N = 3199460.271. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.2

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: IS.DUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core9001800270036004500SE +/- 15.41, N = 34064.081. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.2

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: FT.CUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core20K40K60K80K100KSE +/- 426.91, N = 3102149.251. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.2

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.DUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core16003200480064008000SE +/- 66.24, N = 77273.961. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.2

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.CUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core14002800420056007000SE +/- 96.53, N = 156692.951. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.2

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: CG.CUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core11K22K33K44K55KSE +/- 396.73, N = 1550419.731. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.2

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: BT.CUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core40K80K120K160K200KSE +/- 248.09, N = 3176096.161. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.2

HPL Linpack

HPL is a well known portable Linpack implementation for distributed memory systems. This test profile is testing HPL upstream directly, outside the scope of the HPC Challenge test profile also available through the Phoronix Test Suite (hpcc). The test profile attempts to generate an optimized HPL.dat input file based on the CPU/memory under test. The automated HPL.dat input generation is still being tuned and thus for now this test profile remains "experimental". Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGFLOPS, More Is BetterHPL Linpack 2.3Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core140280420560700SE +/- 0.39, N = 3631.301. (CC) gcc options: -O2 -lopenblas -lm -lmpi

Glibc Benchmarks

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: pthread_onceUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1.27252.5453.81755.096.3625SE +/- 0.00045, N = 35.655471. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: sincosUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core918273645SE +/- 0.01, N = 338.981. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: ffsllUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1.27592.55183.82775.10366.3795SE +/- 0.00011, N = 35.670551. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: atanhUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core714212835SE +/- 0.00, N = 328.181. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: asinhUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core510152025SE +/- 0.00, N = 322.581. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: tanhUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core612182430SE +/- 0.00, N = 326.971. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: sqrtUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core246810SE +/- 0.00536, N = 38.245741. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: sinhUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core510152025SE +/- 0.02, N = 322.961. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: modfUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core246810SE +/- 0.00034, N = 36.619571. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: log2Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core3691215SE +/- 0.00, N = 310.421. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: sinUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1428425670SE +/- 0.00, N = 363.121. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: powUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core816243240SE +/- 0.13, N = 335.561. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: ffsUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1.27582.55163.82745.10326.379SE +/- 0.00021, N = 35.670321. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: expUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core48121620SE +/- 0.00, N = 315.141. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: cosUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core1632486480SE +/- 0.01, N = 370.821. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Compile Bench

Compilebench tries to age a filesystem by simulating some of the disk IO common in creating, compiling, patching, stating and reading kernel trees. It indirectly measures how well filesystems can maintain directory locality as the disk fills up and directories age. This current test is setup to use the makej mode with 10 initial directories Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterCompile Bench 0.6Test: Read Compiled TreeUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core6001200180024003000SE +/- 6.72, N = 32657.42

OpenBenchmarking.orgMB/s, More Is BetterCompile Bench 0.6Test: Initial CreateUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core90180270360450SE +/- 4.73, N = 3399.79

OpenBenchmarking.orgMB/s, More Is BetterCompile Bench 0.6Test: CompileUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core30060090012001500SE +/- 11.78, N = 31613.38

Llama.cpp

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b4397Backend: CPU BLAS - Model: granite-3.0-3b-a800m-instruct-Q8_0 - Test: Prompt Processing 512Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core4080120160200SE +/- 6.45, N = 12193.811. (CXX) g++ options: -O3

Scikit-Learn

Scikit-learn is a Python module for machine learning built on NumPy, SciPy, and is BSD-licensed. Learn more via the OpenBenchmarking.org test page.

Benchmark: Sparse Random Projections / 100 Iterations

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Kernel PCA Solvers / Time vs. N Components

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Kernel PCA Solvers / Time vs. N Samples

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Hist Gradient Boosting Categorical Only

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Plot Non-Negative Matrix Factorization

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Plot Polynomial Kernel Approximation

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: 20 Newsgroups / Logistic Regression

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Hist Gradient Boosting Higgs Boson

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Plot Singular Value Decomposition

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/libblas.so.3: undefined symbol: gotoblas

Benchmark: Hist Gradient Boosting Threading

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Isotonic / Perturbed Logarithm

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Hist Gradient Boosting Adult

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Covertype Dataset Benchmark

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Sample Without Replacement

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: RCV1 Logreg Convergencet

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Isotonic / Pathological

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Plot Parallel Pairwise

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Hist Gradient Boosting

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Plot Incremental PCA

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Isotonic / Logistic

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: TSNE MNIST Dataset

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: LocalOutlierFactor

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Feature Expansions

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Plot OMP vs. LARS

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Plot Hierarchical

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Text Vectorizers

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Plot Fast KMeans

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Isolation Forest

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Plot Lasso Path

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: SGDOneClassSVM

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: SGD Regression

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Plot Neighbors

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: MNIST Dataset

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Plot Ward

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Sparsify

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Glmnet

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Lasso

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: Tree

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: SAGA

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

Benchmark: GLM

Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core: The test quit with a non-zero exit status. E: ImportError: /lib/x86_64-linux-gnu/liblapack.so.3: undefined symbol: gotoblas

NCNN

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: Vulkan GPU - Model: FastestDetUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core48121620SE +/- 0.64, N = 317.29MIN: 15.96 / MAX: 23.341. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU - Model: FastestDetUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core48121620SE +/- 0.32, N = 1217.79MIN: 16.23 / MAX: 24.591. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPUv2-yolov3v2-yolov3 - Model: mobilenetv2-yolov3Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core612182430SE +/- 0.42, N = 1223.31MIN: 20.72 / MAX: 49.061. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU-v3-v3 - Model: mobilenet-v3Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core48121620SE +/- 1.23, N = 1214.36MIN: 12.51 / MAX: 14521. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20241226Target: CPU - Model: mobilenetUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core612182430SE +/- 0.42, N = 1223.31MIN: 20.72 / MAX: 49.061. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

spaCy

The spaCy library is an open-source solution for advanced neural language processing (NLP). The spaCy library leverages Python and is a leading neural language processing solution. This test profile times the spaCy CPU performance with various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgtokens/sec, More Is BetterspaCy 3.4.1Model: en_core_web_trfUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core6001200180024003000SE +/- 96.13, N = 32751

Stress-NG

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: IO_uringUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core40K80K120K160K200KSE +/- 4811.73, N = 12174744.181. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

OpenBenchmarking.orgBogo Ops/s, More Is BetterStress-NG 0.18.09Test: CryptoUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core90M180M270M360M450MSE +/- 64894044.34, N = 15402622178.281. (CXX) g++ options: -O2 -std=gnu99 -lc -lm

TensorFlow Lite

This is a benchmark of the TensorFlow Lite implementation focused on TensorFlow machine learning for mobile, IoT, edge, and other cases. The current Linux support is limited to running on CPUs. This test profile is measuring the average inference time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: Inception ResNet V2Ubuntu 22.04 - 2 x AMD EPYC 9274F 24-Core15K30K45K60K75KSE +/- 1796.42, N = 1568557.9

LiteRT

OpenBenchmarking.orgMicroseconds, Fewer Is BetterLiteRT 2024-10-15Model: Mobilenet QuantUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core7001400210028003500SE +/- 51.59, N = 153255.76

Intel MPI Benchmarks

Intel MPI Benchmarks for stressing MPI implementations. At this point the test profile aggregates results for some common MPI functionality. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgAverage Mbytes/sec, More Is BetterIntel MPI Benchmarks 2019.3Test: IMB-MPI1 PingPongUbuntu 22.04 - 2 x AMD EPYC 9274F 24-Core8001600240032004000SE +/- 61.28, N = 153739.53MIN: 3.8 / MAX: 14414.31. (CXX) g++ options: -O0 -pedantic -fopenmp -lmpi_cxx -lmpi

244 Results Shown

Llamafile:
  wizardcoder-python-34b-v1.0.Q6_K - Prompt Processing 2048
  wizardcoder-python-34b-v1.0.Q6_K - Prompt Processing 1024
  wizardcoder-python-34b-v1.0.Q6_K - Prompt Processing 512
  wizardcoder-python-34b-v1.0.Q6_K - Prompt Processing 256
  mistral-7b-instruct-v0.2.Q5_K_M - Prompt Processing 2048
  mistral-7b-instruct-v0.2.Q5_K_M - Prompt Processing 1024
  mistral-7b-instruct-v0.2.Q5_K_M - Prompt Processing 512
  mistral-7b-instruct-v0.2.Q5_K_M - Prompt Processing 256
  wizardcoder-python-34b-v1.0.Q6_K - Text Generation 128
  TinyLlama-1.1B-Chat-v1.0.BF16 - Prompt Processing 2048
  TinyLlama-1.1B-Chat-v1.0.BF16 - Prompt Processing 1024
  wizardcoder-python-34b-v1.0.Q6_K - Text Generation 16
  mistral-7b-instruct-v0.2.Q5_K_M - Text Generation 128
  TinyLlama-1.1B-Chat-v1.0.BF16 - Prompt Processing 512
  TinyLlama-1.1B-Chat-v1.0.BF16 - Prompt Processing 256
  mistral-7b-instruct-v0.2.Q5_K_M - Text Generation 16
  TinyLlama-1.1B-Chat-v1.0.BF16 - Text Generation 128
  Llama-3.2-3B-Instruct.Q6_K - Prompt Processing 2048
  Llama-3.2-3B-Instruct.Q6_K - Prompt Processing 1024
  TinyLlama-1.1B-Chat-v1.0.BF16 - Text Generation 16
  Llama-3.2-3B-Instruct.Q6_K - Prompt Processing 512
  Llama-3.2-3B-Instruct.Q6_K - Prompt Processing 256
  Llama-3.2-3B-Instruct.Q6_K - Text Generation 128
  Llama-3.2-3B-Instruct.Q6_K - Text Generation 16
Llama.cpp:
  NVIDIA CUDA - granite-3.0-3b-a800m-instruct-Q8_0 - Prompt Processing 2048
  NVIDIA CUDA - granite-3.0-3b-a800m-instruct-Q8_0 - Prompt Processing 1024
  NVIDIA CUDA - granite-3.0-3b-a800m-instruct-Q8_0 - Prompt Processing 512
  NVIDIA CUDA - granite-3.0-3b-a800m-instruct-Q8_0 - Text Generation 128
  CPU BLAS - granite-3.0-3b-a800m-instruct-Q8_0 - Prompt Processing 2048
  CPU BLAS - granite-3.0-3b-a800m-instruct-Q8_0 - Prompt Processing 1024
  NVIDIA CUDA - Mistral-7B-Instruct-v0.3-Q8_0 - Prompt Processing 2048
  NVIDIA CUDA - Mistral-7B-Instruct-v0.3-Q8_0 - Prompt Processing 1024
  NVIDIA CUDA - Mistral-7B-Instruct-v0.3-Q8_0 - Prompt Processing 512
  CPU BLAS - granite-3.0-3b-a800m-instruct-Q8_0 - Text Generation 128
  NVIDIA CUDA - Mistral-7B-Instruct-v0.3-Q8_0 - Text Generation 128
  CPU BLAS - Mistral-7B-Instruct-v0.3-Q8_0 - Prompt Processing 2048
  CPU BLAS - Mistral-7B-Instruct-v0.3-Q8_0 - Prompt Processing 1024
  CPU BLAS - Mistral-7B-Instruct-v0.3-Q8_0 - Prompt Processing 512
  NVIDIA CUDA - Llama-3.1-Tulu-3-8B-Q8_0 - Prompt Processing 2048
  NVIDIA CUDA - Llama-3.1-Tulu-3-8B-Q8_0 - Prompt Processing 1024
  NVIDIA CUDA - Llama-3.1-Tulu-3-8B-Q8_0 - Prompt Processing 512
  CPU BLAS - Mistral-7B-Instruct-v0.3-Q8_0 - Text Generation 128
  NVIDIA CUDA - Llama-3.1-Tulu-3-8B-Q8_0 - Text Generation 128
  CPU BLAS - Llama-3.1-Tulu-3-8B-Q8_0 - Prompt Processing 2048
  CPU BLAS - Llama-3.1-Tulu-3-8B-Q8_0 - Prompt Processing 1024
  CPU BLAS - Llama-3.1-Tulu-3-8B-Q8_0 - Prompt Processing 512
  CPU BLAS - Llama-3.1-Tulu-3-8B-Q8_0 - Text Generation 128
AI Benchmark Alpha:
  Device AI Score
  Device Training Score
  Device Inference Score
PyBench
NCNN:
  Vulkan GPU - vision_transformer
  Vulkan GPU - regnety_400m
  Vulkan GPU - squeezenet_ssd
  Vulkan GPU - yolov4-tiny
  Vulkan GPUv2-yolov3v2-yolov3 - mobilenetv2-yolov3
  Vulkan GPU - resnet50
  Vulkan GPU - alexnet
  Vulkan GPU - resnet18
  Vulkan GPU - vgg16
  Vulkan GPU - googlenet
  Vulkan GPU - blazeface
  Vulkan GPU - efficientnet-b0
  Vulkan GPU - mnasnet
  Vulkan GPU - shufflenet-v2
  Vulkan GPU-v3-v3 - mobilenet-v3
  Vulkan GPU-v2-v2 - mobilenet-v2
  Vulkan GPU - mobilenet
  CPU - vision_transformer
  CPU - regnety_400m
  CPU - squeezenet_ssd
  CPU - yolov4-tiny
  CPU - resnet50
  CPU - alexnet
  CPU - resnet18
  CPU - vgg16
  CPU - googlenet
  CPU - blazeface
  CPU - efficientnet-b0
  CPU - mnasnet
  CPU - shufflenet-v2
  CPU-v2-v2 - mobilenet-v2
spaCy
Stress-NG:
  Hyperbolic Trigonometric Math
  POSIX Regular Expressions
  System V Message Passing
  Glibc Qsort Data Sorting
  Glibc C String Functions
  Integer Bit Operations
  Bessel Math Operations
  Vector Floating Point
  Bitonic Integer Sort
  Trigonometric Math
  Fused Multiply-Add
  Radix String Sort
  Fractal Generator
  Context Switching
  Wide Vector Math
  Logarithmic Math
  Jpeg Compression
  Exponential Math
  Socket Activity
  Mixed Scheduler
  Vector Shuffle
  Memory Copying
  Matrix 3D Math
  Floating Point
  x86_64 RdRand
  Function Call
  Integer Math
  AVX-512 VNNI
  Vector Math
  Matrix Math
  Semaphores
  Power Math
  CPU Stress
  CPU Cache
  SENDFILE
  AVL Tree
  Pthread
  Forking
  Cloning
  Malloc
  Atomic
  Mutex
  MEMFD
  Futex
  Zlib
  Poll
  Pipe
  NUMA
  MMAP
  Hash
PyTorch:
  NVIDIA CUDA GPU - 512 - Efficientnet_v2_l
  NVIDIA CUDA GPU - 256 - Efficientnet_v2_l
  NVIDIA CUDA GPU - 64 - Efficientnet_v2_l
  NVIDIA CUDA GPU - 32 - Efficientnet_v2_l
  NVIDIA CUDA GPU - 16 - Efficientnet_v2_l
  NVIDIA CUDA GPU - 1 - Efficientnet_v2_l
  NVIDIA CUDA GPU - 512 - ResNet-152
  NVIDIA CUDA GPU - 256 - ResNet-152
  NVIDIA CUDA GPU - 64 - ResNet-152
  NVIDIA CUDA GPU - 512 - ResNet-50
  NVIDIA CUDA GPU - 32 - ResNet-152
  NVIDIA CUDA GPU - 256 - ResNet-50
  NVIDIA CUDA GPU - 16 - ResNet-152
  NVIDIA CUDA GPU - 64 - ResNet-50
  NVIDIA CUDA GPU - 32 - ResNet-50
  NVIDIA CUDA GPU - 16 - ResNet-50
  NVIDIA CUDA GPU - 1 - ResNet-152
  NVIDIA CUDA GPU - 1 - ResNet-50
  CPU - 512 - Efficientnet_v2_l
  CPU - 256 - Efficientnet_v2_l
  CPU - 64 - Efficientnet_v2_l
  CPU - 32 - Efficientnet_v2_l
  CPU - 16 - Efficientnet_v2_l
  CPU - 1 - Efficientnet_v2_l
  CPU - 512 - ResNet-152
  CPU - 256 - ResNet-152
  CPU - 64 - ResNet-152
  CPU - 512 - ResNet-50
  CPU - 32 - ResNet-152
  CPU - 256 - ResNet-50
  CPU - 16 - ResNet-152
  CPU - 64 - ResNet-50
  CPU - 32 - ResNet-50
  CPU - 16 - ResNet-50
  CPU - 1 - ResNet-152
  CPU - 1 - ResNet-50
TensorFlow Lite:
  Mobilenet Quant
  Mobilenet Float
  NASNet Mobile
  Inception V4
  SqueezeNet
LiteRT:
  Quantized COCO SSD MobileNet v1
  Inception ResNet V2
  Mobilenet Float
  NASNet Mobile
  Inception V4
  SqueezeNet
  DeepLab V3
Intel MPI Benchmarks:
  IMB-MPI1 Sendrecv:
    Average usec
    Average Mbytes/sec
  IMB-MPI1 Exchange:
    Average usec
    Average Mbytes/sec
  IMB-P2P PingPong:
    Average Msg/sec
R Benchmark
Cython Benchmark
Numpy Benchmark
oneDNN:
  Recurrent Neural Network Inference - CPU
  Recurrent Neural Network Training - CPU
  Deconvolution Batch shapes_3d - CPU
  Deconvolution Batch shapes_1d - CPU
  Convolution Batch Shapes Auto - CPU
  IP Shapes 3D - CPU
  IP Shapes 1D - CPU
Timed LLVM Compilation:
  Unix Makefiles
  Ninja
Timed Linux Kernel Compilation:
  allmodconfig
  defconfig
Timed GCC Compilation
Epoch
Timed MrBayes Analysis
NAS Parallel Benchmarks:
  SP.C
  SP.B
  MG.C
  LU.C
  IS.D
  FT.C
  EP.D
  EP.C
  CG.C
  BT.C
HPL Linpack
Glibc Benchmarks:
  pthread_once
  sincos
  ffsll
  atanh
  asinh
  tanh
  sqrt
  sinh
  modf
  log2
  sin
  pow
  ffs
  exp
  cos
Compile Bench:
  Read Compiled Tree
  Initial Create
  Compile
Llama.cpp
NCNN:
  Vulkan GPU - FastestDet
  CPU - FastestDet
  CPUv2-yolov3v2-yolov3 - mobilenetv2-yolov3
  CPU-v3-v3 - mobilenet-v3
  CPU - mobilenet
spaCy
Stress-NG:
  IO_uring
  Crypto
TensorFlow Lite
LiteRT
Intel MPI Benchmarks