n1n1

ARMv8 Neoverse-N1 testing with a GIGABYTE G242-P36-00 MP32-AR2-00 v01000100 (F31k SCP: 2.10.20220531 BIOS) and ASPEED on Ubuntu 23.10 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2403174-NE-N1N13670960
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results

Limit displaying results to tests within:

C/C++ Compiler Tests 2 Tests
CPU Massive 6 Tests
Creator Workloads 7 Tests
Encoding 2 Tests
HPC - High Performance Computing 3 Tests
Imaging 2 Tests
Machine Learning 3 Tests
Multi-Core 7 Tests
Intel oneAPI 2 Tests
Python Tests 2 Tests
Server CPU Tests 4 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
a
March 17
  15 Minutes
aa
March 17
  7 Hours, 43 Minutes
b
March 17
  2 Hours, 32 Minutes
c
March 17
  2 Hours, 15 Minutes
Invert Hiding All Results Option
  3 Hours, 11 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


n1n1 Suite 1.0.0 System Test suite extracted from n1n1 . pts/jpegxl-1.6.0 sample-4.png out.jxl -q 80 --num_reps 80 Input: PNG - Quality: 80 pts/jpegxl-1.6.0 sample-4.png out.jxl -q 90 --num_reps 50 Input: PNG - Quality: 90 pts/jpegxl-1.6.0 --lossless_jpeg=0 sample-photo-6000x4000.JPG out.jxl -q 80 --num_reps 80 Input: JPEG - Quality: 80 pts/jpegxl-1.6.0 --lossless_jpeg=0 sample-photo-6000x4000.JPG out.jxl -q 90 --num_reps 50 Input: JPEG - Quality: 90 pts/jpegxl-1.6.0 sample-4.png out.jxl -q 100 --num_reps 20 Input: PNG - Quality: 100 pts/jpegxl-1.6.0 --lossless_jpeg=0 sample-photo-6000x4000.JPG out.jxl -q 100 --num_reps 20 Input: JPEG - Quality: 100 pts/jpegxl-decode-1.6.0 --num_threads=1 --num_reps=90 CPU Threads: 1 pts/jpegxl-decode-1.6.0 --num_reps=250 CPU Threads: All pts/srsran-2.2.0 tests/benchmarks/phy/upper/channel_processors/pdsch_processor_benchmark -m throughput_total -R 350 -B 10 -P 4port_4layer_scs30_100MHz_256qam Test: PDSCH Processor Benchmark, Throughput Total pts/srsran-2.2.0 tests/benchmarks/phy/upper/channel_processors/pusch/pusch_processor_benchmark -m throughput_total -R 100 -B 10 -P pusch_scs30_100MHz_256qam_max Test: PUSCH Processor Benchmark, Throughput Total pts/srsran-2.2.0 tests/benchmarks/phy/upper/channel_processors/pdsch_processor_benchmark -m throughput_thread -R 350 -B 10 -T 1 -P 4port_4layer_scs30_100MHz_256qam Test: PDSCH Processor Benchmark, Throughput Thread pts/srsran-2.2.0 tests/benchmarks/phy/upper/channel_processors/pusch/pusch_processor_benchmark -m throughput_thread -R 350 -B 10 -T 1 -t 0 -P pusch_scs30_100MHz_256qam_max Test: PUSCH Processor Benchmark, Throughput Thread pts/svt-av1-2.12.0 --preset 4 -n 160 -i Bosphorus_3840x2160.y4m -w 3840 -h 2160 Encoder Mode: Preset 4 - Input: Bosphorus 4K pts/svt-av1-2.12.0 --preset 8 -i Bosphorus_3840x2160.y4m -w 3840 -h 2160 Encoder Mode: Preset 8 - Input: Bosphorus 4K pts/svt-av1-2.12.0 --preset 12 -i Bosphorus_3840x2160.y4m -w 3840 -h 2160 Encoder Mode: Preset 12 - Input: Bosphorus 4K pts/svt-av1-2.12.0 --preset 13 -i Bosphorus_3840x2160.y4m -w 3840 -h 2160 Encoder Mode: Preset 13 - Input: Bosphorus 4K pts/svt-av1-2.12.0 --preset 4 -n 160 -i Bosphorus_1920x1080_120fps_420_8bit_YUV.yuv -w 1920 -h 1080 Encoder Mode: Preset 4 - Input: Bosphorus 1080p pts/svt-av1-2.12.0 --preset 8 -i Bosphorus_1920x1080_120fps_420_8bit_YUV.yuv -w 1920 -h 1080 Encoder Mode: Preset 8 - Input: Bosphorus 1080p pts/svt-av1-2.12.0 --preset 12 -i Bosphorus_1920x1080_120fps_420_8bit_YUV.yuv -w 1920 -h 1080 Encoder Mode: Preset 12 - Input: Bosphorus 1080p pts/svt-av1-2.12.0 --preset 13 -i Bosphorus_1920x1080_120fps_420_8bit_YUV.yuv -w 1920 -h 1080 Encoder Mode: Preset 13 - Input: Bosphorus 1080p pts/stockfish-1.5.0 Chess Benchmark pts/build-linux-kernel-1.16.0 defconfig Build: defconfig pts/build-linux-kernel-1.16.0 allmodconfig Build: allmodconfig pts/compress-pbzip2-1.6.1 FreeBSD-13.0-RELEASE-amd64-memstick.img Compression pts/primesieve-1.10.0 1e12 Length: 1e12 pts/primesieve-1.10.0 1e13 Length: 1e13 pts/onednn-3.4.0 --ip --batch=inputs/ip/shapes_1d --engine=cpu Harness: IP Shapes 1D - Engine: CPU pts/onednn-3.4.0 --ip --batch=inputs/ip/shapes_3d --engine=cpu Harness: IP Shapes 3D - Engine: CPU pts/onednn-3.4.0 --conv --batch=inputs/conv/shapes_auto --engine=cpu Harness: Convolution Batch Shapes Auto - Engine: CPU pts/onednn-3.4.0 --deconv --batch=inputs/deconv/shapes_1d --engine=cpu Harness: Deconvolution Batch shapes_1d - Engine: CPU pts/onednn-3.4.0 --deconv --batch=inputs/deconv/shapes_3d --engine=cpu Harness: Deconvolution Batch shapes_3d - Engine: CPU pts/onednn-3.4.0 --rnn --batch=inputs/rnn/perf_rnn_training --engine=cpu Harness: Recurrent Neural Network Training - Engine: CPU pts/onednn-3.4.0 --rnn --batch=inputs/rnn/perf_rnn_inference_lb --engine=cpu Harness: Recurrent Neural Network Inference - Engine: CPU pts/deepsparse-1.7.0 zoo:nlp/document_classification/obert-base/pytorch/huggingface/imdb/base-none --scenario async Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream pts/deepsparse-1.7.0 zoo:nlp/document_classification/obert-base/pytorch/huggingface/imdb/base-none --scenario sync Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Synchronous Single-Stream pts/deepsparse-1.7.0 zoo:nlp/sentiment_analysis/oberta-base/pytorch/huggingface/sst2/pruned90_quant-none --input_shapes='[1,128]' --scenario async Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream pts/deepsparse-1.7.0 zoo:nlp/sentiment_analysis/oberta-base/pytorch/huggingface/sst2/pruned90_quant-none --input_shapes='[1,128]' --scenario sync Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Synchronous Single-Stream pts/deepsparse-1.7.0 zoo:cv/classification/resnet_v1-50/pytorch/sparseml/imagenet/base-none --scenario async Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream pts/deepsparse-1.7.0 zoo:cv/classification/resnet_v1-50/pytorch/sparseml/imagenet/base-none --scenario sync Model: ResNet-50, Baseline - Scenario: Synchronous Single-Stream pts/deepsparse-1.7.0 zoo:cv/classification/resnet_v1-50/pytorch/sparseml/imagenet/pruned95_uniform_quant-none --scenario async Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream pts/deepsparse-1.7.0 zoo:cv/classification/resnet_v1-50/pytorch/sparseml/imagenet/pruned95_uniform_quant-none --scenario sync Model: ResNet-50, Sparse INT8 - Scenario: Synchronous Single-Stream pts/deepsparse-1.7.0 zoo:llama2-7b-llama2_chat_llama2_pretrain-base_quantized --scenario async Model: Llama2 Chat 7b Quantized - Scenario: Asynchronous Multi-Stream pts/deepsparse-1.7.0 zoo:llama2-7b-llama2_chat_llama2_pretrain-base_quantized --scenario sync Model: Llama2 Chat 7b Quantized - Scenario: Synchronous Single-Stream pts/deepsparse-1.7.0 zoo:cv/classification/resnet_v1-50/pytorch/sparseml/imagenet/base-none --scenario async Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream pts/deepsparse-1.7.0 zoo:cv/classification/resnet_v1-50/pytorch/sparseml/imagenet/base-none --scenario sync Model: CV Classification, ResNet-50 ImageNet - Scenario: Synchronous Single-Stream pts/deepsparse-1.7.0 zoo:cv/detection/yolov5-s/pytorch/ultralytics/coco/pruned85-none --scenario async Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream pts/deepsparse-1.7.0 zoo:cv/detection/yolov5-s/pytorch/ultralytics/coco/pruned85-none --scenario sync Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Synchronous Single-Stream pts/deepsparse-1.7.0 zoo:nlp/text_classification/distilbert-none/pytorch/huggingface/mnli/base-none --scenario async Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream pts/deepsparse-1.7.0 zoo:nlp/text_classification/distilbert-none/pytorch/huggingface/mnli/base-none --scenario sync Model: NLP Text Classification, DistilBERT mnli - Scenario: Synchronous Single-Stream pts/deepsparse-1.7.0 zoo:cv/segmentation/yolact-darknet53/pytorch/dbolya/coco/pruned90-none --scenario async Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream pts/deepsparse-1.7.0 zoo:cv/segmentation/yolact-darknet53/pytorch/dbolya/coco/pruned90-none --scenario sync Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Synchronous Single-Stream pts/deepsparse-1.7.0 zoo:nlp/question_answering/obert-large/pytorch/huggingface/squad/pruned97_quant-none --input_shapes='[1,128]' --scenario async Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream pts/deepsparse-1.7.0 zoo:nlp/question_answering/obert-large/pytorch/huggingface/squad/pruned97_quant-none --input_shapes='[1,128]' --scenario sync Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Synchronous Single-Stream pts/deepsparse-1.7.0 zoo:nlp/token_classification/bert-base/pytorch/huggingface/conll2003/base-none --scenario async Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream pts/deepsparse-1.7.0 zoo:nlp/token_classification/bert-base/pytorch/huggingface/conll2003/base-none --scenario sync Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Synchronous Single-Stream pts/draco-1.6.1 -i lion.ply Model: Lion pts/draco-1.6.1 -i church.ply Model: Church Facade pts/openvino-1.5.0 -m models/intel/face-detection-0206/FP16/face-detection-0206.xml -d CPU Model: Face Detection FP16 - Device: CPU pts/openvino-1.5.0 -m models/intel/person-detection-0303/FP16/person-detection-0303.xml -d CPU Model: Person Detection FP16 - Device: CPU pts/openvino-1.5.0 -m models/intel/person-detection-0303/FP32/person-detection-0303.xml -d CPU Model: Person Detection FP32 - Device: CPU pts/openvino-1.5.0 -m models/intel/vehicle-detection-0202/FP16/vehicle-detection-0202.xml -d CPU Model: Vehicle Detection FP16 - Device: CPU pts/openvino-1.5.0 -m models/intel/face-detection-0206/FP16-INT8/face-detection-0206.xml -d CPU Model: Face Detection FP16-INT8 - Device: CPU pts/openvino-1.5.0 -m models/intel/face-detection-retail-0005/FP16/face-detection-retail-0005.xml -d CPU Model: Face Detection Retail FP16 - Device: CPU pts/openvino-1.5.0 -m models/intel/road-segmentation-adas-0001/FP16/road-segmentation-adas-0001.xml -d CPU Model: Road Segmentation ADAS FP16 - Device: CPU pts/openvino-1.5.0 -m models/intel/vehicle-detection-0202/FP16-INT8/vehicle-detection-0202.xml -d CPU Model: Vehicle Detection FP16-INT8 - Device: CPU pts/openvino-1.5.0 -m models/intel/weld-porosity-detection-0001/FP16/weld-porosity-detection-0001.xml -d CPU Model: Weld Porosity Detection FP16 - Device: CPU pts/openvino-1.5.0 -m models/intel/face-detection-retail-0005/FP16-INT8/face-detection-retail-0005.xml -d CPU Model: Face Detection Retail FP16-INT8 - Device: CPU pts/openvino-1.5.0 -m models/intel/road-segmentation-adas-0001/FP16-INT8/road-segmentation-adas-0001.xml -d CPU Model: Road Segmentation ADAS FP16-INT8 - Device: CPU pts/openvino-1.5.0 -m models/intel/machine-translation-nar-en-de-0002/FP16/machine-translation-nar-en-de-0002.xml -d CPU Model: Machine Translation EN To DE FP16 - Device: CPU pts/openvino-1.5.0 -m models/intel/weld-porosity-detection-0001/FP16-INT8/weld-porosity-detection-0001.xml -d CPU Model: Weld Porosity Detection FP16-INT8 - Device: CPU pts/openvino-1.5.0 -m models/intel/person-vehicle-bike-detection-2004/FP16/person-vehicle-bike-detection-2004.xml -d CPU Model: Person Vehicle Bike Detection FP16 - Device: CPU pts/openvino-1.5.0 -m models/intel/noise-suppression-poconetlike-0001/FP16/noise-suppression-poconetlike-0001.xml -d CPU Model: Noise Suppression Poconet-Like FP16 - Device: CPU pts/openvino-1.5.0 -m models/intel/handwritten-english-recognition-0001/FP16/handwritten-english-recognition-0001.xml -d CPU Model: Handwritten English Recognition FP16 - Device: CPU pts/openvino-1.5.0 -m models/intel/person-reidentification-retail-0277/FP16/person-reidentification-retail-0277.xml -d CPU Model: Person Re-Identification Retail FP16 - Device: CPU pts/openvino-1.5.0 -m models/intel/age-gender-recognition-retail-0013/FP16/age-gender-recognition-retail-0013.xml -d CPU Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU pts/openvino-1.5.0 -m models/intel/handwritten-english-recognition-0001/FP16-INT8/handwritten-english-recognition-0001.xml -d CPU Model: Handwritten English Recognition FP16-INT8 - Device: CPU pts/openvino-1.5.0 -m models/intel/age-gender-recognition-retail-0013/FP16-INT8/age-gender-recognition-retail-0013.xml -d CPU Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU pts/encode-wavpack-1.5.0 WAV To WavPack