eoy2024
AMD EPYC 4564P 16-Core testing with a Supermicro AS-3015A-I H13SAE-MF v1.00 (2.1 BIOS) and ASPEED on Ubuntu 24.04 via the Phoronix Test Suite.
HTML result view exported from: https://openbenchmarking.org/result/2412061-NE-EOY20243073&grs.
LiteRT
Model: Quantized COCO SSD MobileNet v1
LiteRT
Model: NASNet Mobile
LiteRT
Model: DeepLab V3
LiteRT
Model: Mobilenet Quant
CP2K Molecular Dynamics
Input: Fayalite-FIST
Llama.cpp
Backend: CPU BLAS - Model: Mistral-7B-Instruct-v0.3-Q8_0 - Test: Prompt Processing 512
Stockfish
Chess Benchmark
RELION
Test: Basic - Device: CPU
LiteRT
Model: Inception V4
Llama.cpp
Backend: CPU BLAS - Model: granite-3.0-3b-a800m-instruct-Q8_0 - Test: Prompt Processing 1024
Renaissance
Test: Apache Spark Bayes
Llama.cpp
Backend: CPU BLAS - Model: Llama-3.1-Tulu-3-8B-Q8_0 - Test: Prompt Processing 512
Llama.cpp
Backend: CPU BLAS - Model: Llama-3.1-Tulu-3-8B-Q8_0 - Test: Prompt Processing 1024
LiteRT
Model: Mobilenet Float
Renaissance
Test: In-Memory Database Shootout
Renaissance
Test: Scala Dotty
CP2K Molecular Dynamics
Input: H20-256
Renaissance
Test: Random Forest
Llama.cpp
Backend: CPU BLAS - Model: Llama-3.1-Tulu-3-8B-Q8_0 - Test: Prompt Processing 2048
Rustls
Benchmark: handshake - Suite: TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384
simdjson
Throughput Test: LargeRandom
Gcrypt Library
XNNPACK
Model: FP16MobileNetV2
oneDNN
Harness: Deconvolution Batch shapes_1d - Engine: CPU
LiteRT
Model: Inception ResNet V2
XNNPACK
Model: FP32MobileNetV2
simdjson
Throughput Test: Kostya
PyPerformance
Benchmark: asyncio_tcp_ssl
Llamafile
Model: Llama-3.2-3B-Instruct.Q6_K - Test: Text Generation 16
XNNPACK
Model: FP32MobileNetV3Large
LiteRT
Model: SqueezeNet
Llama.cpp
Backend: CPU BLAS - Model: Mistral-7B-Instruct-v0.3-Q8_0 - Test: Prompt Processing 2048
Renaissance
Test: Genetic Algorithm Using Jenetics + Futures
XNNPACK
Model: FP16MobileNetV3Large
simdjson
Throughput Test: TopTweet
Llama.cpp
Backend: CPU BLAS - Model: granite-3.0-3b-a800m-instruct-Q8_0 - Test: Text Generation 128
XNNPACK
Model: FP32MobileNetV1
Renaissance
Test: Gaussian Mixture Model
XNNPACK
Model: FP16MobileNetV1
XNNPACK
Model: FP32MobileNetV3Small
Renaissance
Test: Savina Reactors.IO
Renaissance
Test: Akka Unbalanced Cobwebbed Tree
SVT-AV1
Encoder Mode: Preset 8 - Input: Bosphorus 1080p
SVT-AV1
Encoder Mode: Preset 8 - Input: Bosphorus 4K
oneDNN
Harness: IP Shapes 3D - Engine: CPU
Renaissance
Test: Finagle HTTP Requests
oneDNN
Harness: IP Shapes 1D - Engine: CPU
Llama.cpp
Backend: CPU BLAS - Model: granite-3.0-3b-a800m-instruct-Q8_0 - Test: Prompt Processing 2048
CP2K Molecular Dynamics
Input: H20-64
oneDNN
Harness: Convolution Batch Shapes Auto - Engine: CPU
x265
Video Input: Bosphorus 4K
SVT-AV1
Encoder Mode: Preset 13 - Input: Bosphorus 1080p
SVT-AV1
Encoder Mode: Preset 5 - Input: Beauty 4K 10-bit
Timed Eigen Compilation
Time To Compile
oneDNN
Harness: Deconvolution Batch shapes_3d - Engine: CPU
Rustls
Benchmark: handshake-resume - Suite: TLS13_CHACHA20_POLY1305_SHA256
Whisper.cpp
Model: ggml-small.en - Input: 2016 State of the Union
Rustls
Benchmark: handshake-ticket - Suite: TLS13_CHACHA20_POLY1305_SHA256
ONNX Runtime
Model: bertsquad-12 - Device: CPU - Executor: Standard
Llamafile
Model: TinyLlama-1.1B-Chat-v1.0.BF16 - Test: Text Generation 128
Rustls
Benchmark: handshake - Suite: TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256
Rustls
Benchmark: handshake-resume - Suite: TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256
Stockfish
Chess Benchmark
POV-Ray
Trace Time
Llamafile
Model: mistral-7b-instruct-v0.2.Q5_K_M - Test: Text Generation 128
Llamafile
Model: Llama-3.2-3B-Instruct.Q6_K - Test: Text Generation 128
Renaissance
Test: ALS Movie Lens
SVT-AV1
Encoder Mode: Preset 13 - Input: Bosphorus 4K
oneDNN
Harness: Recurrent Neural Network Inference - Engine: CPU
ONNX Runtime
Model: ZFNet-512 - Device: CPU - Executor: Standard
x265
Video Input: Bosphorus 1080p
ONNX Runtime
Model: fcn-resnet101-11 - Device: CPU - Executor: Standard
simdjson
Throughput Test: PartialTweets
Whisperfile
Model Size: Small
ONNX Runtime
Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Standard
NAMD
Input: ATPase with 327,506 Atoms
ONNX Runtime
Model: ResNet101_DUC_HDC-12 - Device: CPU - Executor: Standard
Apache CouchDB
Bulk Size: 100 - Inserts: 3000 - Rounds: 30
Numpy Benchmark
PyPerformance
Benchmark: chaos
ACES DGEMM
Sustained Floating-Point Rate
SVT-AV1
Encoder Mode: Preset 5 - Input: Bosphorus 1080p
PyPerformance
Benchmark: float
XNNPACK
Model: FP16MobileNetV3Small
Rustls
Benchmark: handshake-ticket - Suite: TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256
XNNPACK
Model: QS8MobileNetV2
Whisperfile
Model Size: Tiny
ONNX Runtime
Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard
Renaissance
Test: Apache Spark PageRank
SVT-AV1
Encoder Mode: Preset 8 - Input: Beauty 4K 10-bit
PyPerformance
Benchmark: raytrace
Rustls
Benchmark: handshake-ticket - Suite: TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384
FinanceBench
Benchmark: Bonds OpenMP
ONNX Runtime
Model: CaffeNet 12-int8 - Device: CPU - Executor: Standard
Apache CouchDB
Bulk Size: 500 - Inserts: 3000 - Rounds: 30
ONNX Runtime
Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Standard
PyPerformance
Benchmark: go
SVT-AV1
Encoder Mode: Preset 3 - Input: Bosphorus 4K
Apache CouchDB
Bulk Size: 300 - Inserts: 1000 - Rounds: 30
Llama.cpp
Backend: CPU BLAS - Model: granite-3.0-3b-a800m-instruct-Q8_0 - Test: Prompt Processing 512
Blender
Blend File: Junkshop - Compute: CPU-Only
SVT-AV1
Encoder Mode: Preset 5 - Input: Bosphorus 4K
OSPRay
Benchmark: gravity_spheres_volume/dim_512/ao/real_time
ONNX Runtime
Model: yolov4 - Device: CPU - Executor: Standard
7-Zip Compression
Test: Decompression Rating
oneDNN
Harness: Recurrent Neural Network Training - Engine: CPU
OpenVINO GenAI
Model: Gemma-7b-int4-ov - Device: CPU
SVT-AV1
Encoder Mode: Preset 13 - Input: Beauty 4K 10-bit
OSPRay
Benchmark: gravity_spheres_volume/dim_512/scivis/real_time
SVT-AV1
Encoder Mode: Preset 3 - Input: Beauty 4K 10-bit
GROMACS
Input: water_GMX50_bare
simdjson
Throughput Test: DistinctUserID
Blender
Blend File: Classroom - Compute: CPU-Only
PyPerformance
Benchmark: pathlib
Llama.cpp
Backend: CPU BLAS - Model: Mistral-7B-Instruct-v0.3-Q8_0 - Test: Text Generation 128
ASTC Encoder
Preset: Exhaustive
ASTC Encoder
Preset: Thorough
Etcpak
Benchmark: Multi-Threaded - Configuration: ETC2
PyPerformance
Benchmark: nbody
SVT-AV1
Encoder Mode: Preset 3 - Input: Bosphorus 1080p
Llama.cpp
Backend: CPU BLAS - Model: Mistral-7B-Instruct-v0.3-Q8_0 - Test: Prompt Processing 1024
Apache CouchDB
Bulk Size: 500 - Inserts: 1000 - Rounds: 30
ASTC Encoder
Preset: Medium
Blender
Blend File: Barbershop - Compute: CPU-Only
PyPerformance
Benchmark: pickle_pure_python
ASTC Encoder
Preset: Very Thorough
PyPerformance
Benchmark: gc_collect
OSPRay
Benchmark: particle_volume/scivis/real_time
Primesieve
Length: 1e13
PyPerformance
Benchmark: regex_compile
Llamafile
Model: wizardcoder-python-34b-v1.0.Q6_K - Test: Text Generation 16
OSPRay
Benchmark: particle_volume/pathtracer/real_time
PyPerformance
Benchmark: async_tree_io
Llamafile
Model: wizardcoder-python-34b-v1.0.Q6_K - Test: Text Generation 128
Blender
Blend File: Fishy Cat - Compute: CPU-Only
Primesieve
Length: 1e12
Rustls
Benchmark: handshake - Suite: TLS13_CHACHA20_POLY1305_SHA256
FinanceBench
Benchmark: Repo OpenMP
PyPerformance
Benchmark: django_template
OSPRay
Benchmark: particle_volume/ao/real_time
BYTE Unix Benchmark
Computational Test: Dhrystone 2
QuantLib
Size: XXS
Y-Cruncher
Pi Digits To Calculate: 1B
Llama.cpp
Backend: CPU BLAS - Model: Llama-3.1-Tulu-3-8B-Q8_0 - Test: Text Generation 128
OpenVINO GenAI
Model: Phi-3-mini-128k-instruct-int4-ov - Device: CPU
Llamafile
Model: TinyLlama-1.1B-Chat-v1.0.BF16 - Test: Text Generation 16
Whisperfile
Model Size: Medium
BYTE Unix Benchmark
Computational Test: Pipe
Llamafile
Model: mistral-7b-instruct-v0.2.Q5_K_M - Test: Text Generation 16
Blender
Blend File: BMW27 - Compute: CPU-Only
OpenSSL
Algorithm: AES-128-GCM
OpenSSL
Algorithm: AES-256-GCM
PyPerformance
Benchmark: python_startup
OSPRay
Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_time
Whisper.cpp
Model: ggml-medium.en - Input: 2016 State of the Union
PyPerformance
Benchmark: asyncio_websockets
OpenVINO GenAI
Model: Falcon-7b-instruct-int4-ov - Device: CPU
QuantLib
Size: S
PyPerformance
Benchmark: xml_etree
7-Zip Compression
Test: Compression Rating
Apache CouchDB
Bulk Size: 100 - Inserts: 1000 - Rounds: 30
Build2
Time To Compile
BYTE Unix Benchmark
Computational Test: System Call
Y-Cruncher
Pi Digits To Calculate: 500M
Whisper.cpp
Model: ggml-base.en - Input: 2016 State of the Union
ONNX Runtime
Model: T5 Encoder - Device: CPU - Executor: Standard
PyPerformance
Benchmark: crypto_pyaes
NAMD
Input: STMV with 1,066,628 Atoms
Apache CouchDB
Bulk Size: 300 - Inserts: 3000 - Rounds: 30
ONNX Runtime
Model: GPT-2 - Device: CPU - Executor: Standard
OpenSSL
Algorithm: ChaCha20-Poly1305
OpenSSL
Algorithm: ChaCha20
BYTE Unix Benchmark
Computational Test: Whetstone Double
ONNX Runtime
Model: super-resolution-10 - Device: CPU - Executor: Standard
Blender
Blend File: Pabellon Barcelona - Compute: CPU-Only
ASTC Encoder
Preset: Fast
Rustls
Benchmark: handshake-resume - Suite: TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384
Apache Cassandra
Test: Writes
Llamafile
Model: wizardcoder-python-34b-v1.0.Q6_K - Test: Prompt Processing 2048
Llamafile
Model: wizardcoder-python-34b-v1.0.Q6_K - Test: Prompt Processing 1024
Llamafile
Model: wizardcoder-python-34b-v1.0.Q6_K - Test: Prompt Processing 512
Llamafile
Model: wizardcoder-python-34b-v1.0.Q6_K - Test: Prompt Processing 256
Llamafile
Model: mistral-7b-instruct-v0.2.Q5_K_M - Test: Prompt Processing 2048
Llamafile
Model: mistral-7b-instruct-v0.2.Q5_K_M - Test: Prompt Processing 1024
Llamafile
Model: mistral-7b-instruct-v0.2.Q5_K_M - Test: Prompt Processing 512
Llamafile
Model: mistral-7b-instruct-v0.2.Q5_K_M - Test: Prompt Processing 256
Llamafile
Model: TinyLlama-1.1B-Chat-v1.0.BF16 - Test: Prompt Processing 2048
Llamafile
Model: TinyLlama-1.1B-Chat-v1.0.BF16 - Test: Prompt Processing 1024
Llamafile
Model: TinyLlama-1.1B-Chat-v1.0.BF16 - Test: Prompt Processing 512
Llamafile
Model: TinyLlama-1.1B-Chat-v1.0.BF16 - Test: Prompt Processing 256
Llamafile
Model: Llama-3.2-3B-Instruct.Q6_K - Test: Prompt Processing 2048
Llamafile
Model: Llama-3.2-3B-Instruct.Q6_K - Test: Prompt Processing 1024
Llamafile
Model: Llama-3.2-3B-Instruct.Q6_K - Test: Prompt Processing 512
Llamafile
Model: Llama-3.2-3B-Instruct.Q6_K - Test: Prompt Processing 256
PyPerformance
Benchmark: json_loads
OpenVINO GenAI
Model: Phi-3-mini-128k-instruct-int4-ov - Device: CPU - Time Per Output Token
OpenVINO GenAI
Model: Phi-3-mini-128k-instruct-int4-ov - Device: CPU - Time To First Token
OpenVINO GenAI
Model: Falcon-7b-instruct-int4-ov - Device: CPU - Time Per Output Token
OpenVINO GenAI
Model: Falcon-7b-instruct-int4-ov - Device: CPU - Time To First Token
OpenVINO GenAI
Model: Gemma-7b-int4-ov - Device: CPU - Time Per Output Token
OpenVINO GenAI
Model: Gemma-7b-int4-ov - Device: CPU - Time To First Token
ONNX Runtime
Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Standard
ONNX Runtime
Model: ResNet101_DUC_HDC-12 - Device: CPU - Executor: Standard
ONNX Runtime
Model: super-resolution-10 - Device: CPU - Executor: Standard
ONNX Runtime
Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Standard
ONNX Runtime
Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard
ONNX Runtime
Model: fcn-resnet101-11 - Device: CPU - Executor: Standard
ONNX Runtime
Model: CaffeNet 12-int8 - Device: CPU - Executor: Standard
ONNX Runtime
Model: bertsquad-12 - Device: CPU - Executor: Standard
ONNX Runtime
Model: T5 Encoder - Device: CPU - Executor: Standard
ONNX Runtime
Model: ZFNet-512 - Device: CPU - Executor: Standard
ONNX Runtime
Model: yolov4 - Device: CPU - Executor: Standard
ONNX Runtime
Model: GPT-2 - Device: CPU - Executor: Standard
Phoronix Test Suite v10.8.5