new tests may aarch64 Ampere ARMv8 Neoverse-N1 testing with a WIWYNN Mt.Jade (2.03.20210719 SCP: BIOS) and ASPEED on Ubuntu 21.10 via the Phoronix Test Suite. A: Processor: Ampere ARMv8 Neoverse-N1 @ 3.00GHz (256 Cores), Motherboard: WIWYNN Mt.Jade (2.03.20210719 SCP: BIOS), Chipset: Ampere Computing LLC Altra PCI Root Complex A, Memory: 512GB, Disk: 3841GB Micron_9300_MTFDHAL3T8TDP + 960GB SAMSUNG MZ1LB960HAJQ-00007, Graphics: ASPEED, Network: Mellanox MT28908 + Intel I210 OS: Ubuntu 21.10, Kernel: 5.13.0-27-generic (aarch64), Display Server: X Server, Compiler: GCC 11.2.0, File-System: ext4, Screen Resolution: 1024x768 B: Processor: Ampere ARMv8 Neoverse-N1 @ 3.00GHz (256 Cores), Motherboard: WIWYNN Mt.Jade (2.03.20210719 SCP: BIOS), Chipset: Ampere Computing LLC Altra PCI Root Complex A, Memory: 512GB, Disk: 3841GB Micron_9300_MTFDHAL3T8TDP + 960GB SAMSUNG MZ1LB960HAJQ-00007, Graphics: ASPEED, Network: Mellanox MT28908 + Intel I210 OS: Ubuntu 21.10, Kernel: 5.13.0-27-generic (aarch64), Display Server: X Server, Compiler: GCC 11.2.0, File-System: ext4, Screen Resolution: 1024x768 perf-bench Benchmark: Epoll Wait ops/sec > Higher Is Better A . 1064 |================================================================== B . 1119 |===================================================================== perf-bench Benchmark: Futex Hash ops/sec > Higher Is Better A . 326958 |=================================================================== B . 326997 |=================================================================== perf-bench Benchmark: Memcpy 1MB GB/sec > Higher Is Better A . 15.73 |==================================================================== B . 14.68 |=============================================================== perf-bench Benchmark: Memset 1MB GB/sec > Higher Is Better A . 44.75 |==================================================================== B . 44.74 |==================================================================== perf-bench Benchmark: Sched Pipe ops/sec > Higher Is Better A . 200009 |=============================================================== B . 212491 |=================================================================== perf-bench Benchmark: Futex Lock-Pi ops/sec > Higher Is Better A . 45 |===================================================================== B . 46 |======================================================================= perf-bench Benchmark: Syscall Basic ops/sec > Higher Is Better A . 7827961 |================================================================== B . 7850909 |================================================================== libavif avifenc 0.10 Encoder Speed: 0 Seconds < Lower Is Better A . 252.06 |=================================================================== B . 252.64 |=================================================================== libavif avifenc 0.10 Encoder Speed: 2 Seconds < Lower Is Better A . 163.44 |================================================================== B . 164.68 |=================================================================== libavif avifenc 0.10 Encoder Speed: 6 Seconds < Lower Is Better A . 4.641 |==================================================================== B . 4.610 |==================================================================== libavif avifenc 0.10 Encoder Speed: 6, Lossless Seconds < Lower Is Better A . 7.879 |==================================================================== B . 7.847 |==================================================================== libavif avifenc 0.10 Encoder Speed: 10, Lossless Seconds < Lower Is Better A . 6.131 |==================================================================== B . 5.958 |================================================================== oneDNN 2.6 Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU ms < Lower Is Better A . 44.23 |==================================================================== B . 43.73 |=================================================================== oneDNN 2.6 Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU ms < Lower Is Better A . 40.82 |==================================================================== B . 41.00 |==================================================================== oneDNN 2.6 Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU ms < Lower Is Better A . 106.98 |=================================================================== B . 106.21 |=================================================================== oneDNN 2.6 Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU ms < Lower Is Better A . 166.90 |=================================================================== B . 159.30 |================================================================ oneDNN 2.6 Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU ms < Lower Is Better oneDNN 2.6 Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU ms < Lower Is Better oneDNN 2.6 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU ms < Lower Is Better A . 20.85 |==================================================================== B . 11.51 |====================================== oneDNN 2.6 Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU ms < Lower Is Better A . 56.44 |============================================================= B . 63.16 |==================================================================== oneDNN 2.6 Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU ms < Lower Is Better A . 49.77 |================================================================== B . 50.96 |==================================================================== oneDNN 2.6 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU ms < Lower Is Better A . 130.81 |====================================================== B . 163.01 |=================================================================== oneDNN 2.6 Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU ms < Lower Is Better A . 89.07 |================================================================= B . 93.34 |==================================================================== oneDNN 2.6 Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU ms < Lower Is Better A . 36.19 |==================================================================== B . 35.85 |=================================================================== oneDNN 2.6 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU ms < Lower Is Better A . 16357.1 |================================================================ B . 16750.3 |================================================================== oneDNN 2.6 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU ms < Lower Is Better A . 12883.7 |================================================================== B . 11904.0 |============================================================= oneDNN 2.6 Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU ms < Lower Is Better A . 18083 |==================================================================== B . 14239 |====================================================== oneDNN 2.6 Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU ms < Lower Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU ms < Lower Is Better oneDNN 2.6 Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU ms < Lower Is Better oneDNN 2.6 Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU ms < Lower Is Better A . 13539.1 |================================================================== B . 10490.5 |=================================================== oneDNN 2.6 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU ms < Lower Is Better A . 34.42 |============================================================== B . 37.94 |==================================================================== oneDNN 2.6 Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU ms < Lower Is Better A . 17223.0 |================================================================== B . 13968.6 |====================================================== oneDNN 2.6 Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU ms < Lower Is Better A . 11352.0 |=========================================================== B . 12769.6 |================================================================== oneDNN 2.6 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU ms < Lower Is Better A . 109.23 |=========================================== B . 168.79 |=================================================================== oneDNN 2.6 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU ms < Lower Is Better WebP2 Image Encode 20220422 Encode Settings: Default Seconds < Lower Is Better A . 3.444 |==================================================================== B . 3.423 |==================================================================== WebP2 Image Encode 20220422 Encode Settings: Quality 75, Compression Effort 7 Seconds < Lower Is Better A . 161.27 |=================================================================== B . 161.48 |=================================================================== WebP2 Image Encode 20220422 Encode Settings: Quality 95, Compression Effort 7 Seconds < Lower Is Better A . 318.77 |=================================================================== B . 318.02 |=================================================================== WebP2 Image Encode 20220422 Encode Settings: Quality 100, Compression Effort 5 Seconds < Lower Is Better A . 5.174 |==================================================================== B . 5.178 |==================================================================== WebP2 Image Encode 20220422 Encode Settings: Quality 100, Lossless Compression Seconds < Lower Is Better A . 625.65 |=================================================================== B . 624.94 |=================================================================== Stress-NG 0.14 Test: MMAP Bogo Ops/s > Higher Is Better A . 2051.87 |================================================================== B . 1967.43 |=============================================================== Stress-NG 0.14 Test: NUMA Bogo Ops/s > Higher Is Better A . 32.37 |==================================================================== B . 32.47 |==================================================================== Stress-NG 0.14 Test: Futex Bogo Ops/s > Higher Is Better A . 328915.8 |================================================================= B . 309072.4 |============================================================= Stress-NG 0.14 Test: MEMFD Bogo Ops/s > Higher Is Better A . 1422.26 |================================================================== B . 1427.31 |================================================================== Stress-NG 0.14 Test: Atomic Bogo Ops/s > Higher Is Better A . 1 |======================================================================== B . 1 |======================================================================== Stress-NG 0.14 Test: Crypto Bogo Ops/s > Higher Is Better A . 333734.79 |================================================================ B . 331354.53 |================================================================ Stress-NG 0.14 Test: Malloc Bogo Ops/s > Higher Is Better A . 114773687.93 |============================================================= B . 110879866.19 |=========================================================== Stress-NG 0.14 Test: Forking Bogo Ops/s > Higher Is Better A . 14393.70 |================================================================= B . 13877.64 |=============================================================== Stress-NG 0.14 Test: IO_uring Bogo Ops/s > Higher Is Better A . 1937964.93 |=============================================================== B . 1917001.58 |============================================================== Stress-NG 0.14 Test: SENDFILE Bogo Ops/s > Higher Is Better A . 1979249.09 |=============================================================== B . 1614754.32 |=================================================== Stress-NG 0.14 Test: CPU Cache Bogo Ops/s > Higher Is Better A . 419.89 |=================================================================== B . 393.71 |=============================================================== Stress-NG 0.14 Test: CPU Stress Bogo Ops/s > Higher Is Better A . 65472.66 |================================================================= B . 65456.30 |================================================================= Stress-NG 0.14 Test: Semaphores Bogo Ops/s > Higher Is Better A . 18356640.23 |============================================================== B . 18374889.50 |============================================================== Stress-NG 0.14 Test: Matrix Math Bogo Ops/s > Higher Is Better A . 1198793.00 |=============================================================== B . 1198297.22 |=============================================================== Stress-NG 0.14 Test: Vector Math Bogo Ops/s > Higher Is Better A . 723548.66 |================================================================ B . 723262.70 |================================================================ Stress-NG 0.14 Test: Memory Copying Bogo Ops/s > Higher Is Better A . 9674.49 |================================================================= B . 9837.19 |================================================================== Stress-NG 0.14 Test: Socket Activity Bogo Ops/s > Higher Is Better A . 18596.34 |================================================================= B . 18652.55 |================================================================= Stress-NG 0.14 Test: Context Switching Bogo Ops/s > Higher Is Better A . 31806812.55 |============================================================== B . 31522925.81 |============================================================= Stress-NG 0.14 Test: Glibc C String Functions Bogo Ops/s > Higher Is Better A . 19212257.31 |============================================================== B . 19155740.91 |============================================================== Stress-NG 0.14 Test: Glibc Qsort Data Sorting Bogo Ops/s > Higher Is Better A . 2477.96 |================================================================== B . 2482.09 |================================================================== Stress-NG 0.14 Test: System V Message Passing Bogo Ops/s > Higher Is Better A . 2426671.24 |=============================================================== B . 2444337.47 |=============================================================== Java JMH Throughput Ops/s > Higher Is Better A . 493572978516.11 |========================================================== B . 480270249798.56 |======================================================== nginx 1.21.1 Concurrent Requests: 1 Requests Per Second > Higher Is Better A . 41813.43 |================================================================= B . 41347.93 |================================================================ nginx 1.21.1 Concurrent Requests: 20 Requests Per Second > Higher Is Better A . 83090.65 |================================================================= B . 82864.74 |================================================================= nginx 1.21.1 Concurrent Requests: 100 Requests Per Second > Higher Is Better A . 68830.20 |================================================================= B . 68559.66 |================================================================= nginx 1.21.1 Concurrent Requests: 200 Requests Per Second > Higher Is Better A . 61794.37 |================================================================= B . 62156.77 |================================================================= nginx 1.21.1 Concurrent Requests: 500 Requests Per Second > Higher Is Better A . 63406.60 |================================================================ B . 63986.76 |================================================================= nginx 1.21.1 Concurrent Requests: 1000 Requests Per Second > Higher Is Better A . 65624.22 |================================================================= B . 65932.64 |================================================================= ONNX Runtime 1.11 Model: GPT-2 - Device: CPU - Executor: Parallel Inferences Per Minute > Higher Is Better A . 1596 |==================================================================== B . 1625 |===================================================================== ONNX Runtime 1.11 Model: GPT-2 - Device: CPU - Executor: Standard Inferences Per Minute > Higher Is Better A . 4235 |======================================================= B . 5345 |===================================================================== ONNX Runtime 1.11 Model: yolov4 - Device: CPU - Executor: Parallel Inferences Per Minute > Higher Is Better A . 214 |===================================================================== B . 216 |====================================================================== ONNX Runtime 1.11 Model: yolov4 - Device: CPU - Executor: Standard Inferences Per Minute > Higher Is Better A . 203 |====================================================================== B . 194 |=================================================================== ONNX Runtime 1.11 Model: bertsquad-12 - Device: CPU - Executor: Parallel Inferences Per Minute > Higher Is Better A . 303 |====================================================================== B . 300 |===================================================================== ONNX Runtime 1.11 Model: bertsquad-12 - Device: CPU - Executor: Standard Inferences Per Minute > Higher Is Better A . 516 |====================================================================== B . 505 |===================================================================== ONNX Runtime 1.11 Model: fcn-resnet101-11 - Device: CPU - Executor: Parallel Inferences Per Minute > Higher Is Better A . 40 |======================================================================= B . 40 |======================================================================= ONNX Runtime 1.11 Model: fcn-resnet101-11 - Device: CPU - Executor: Standard Inferences Per Minute > Higher Is Better A . 38 |======================================================================= B . 37 |===================================================================== ONNX Runtime 1.11 Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel Inferences Per Minute > Higher Is Better A . 239 |===================================================================== B . 241 |====================================================================== ONNX Runtime 1.11 Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard Inferences Per Minute > Higher Is Better A . 229 |============================================================ B . 266 |====================================================================== ONNX Runtime 1.11 Model: super-resolution-10 - Device: CPU - Executor: Parallel Inferences Per Minute > Higher Is Better A . 3558 |===================================================================== B . 3555 |===================================================================== ONNX Runtime 1.11 Model: super-resolution-10 - Device: CPU - Executor: Standard Inferences Per Minute > Higher Is Better A . 4242 |==================================================================== B . 4276 |===================================================================== Apache HTTP Server 2.4.48 Concurrent Requests: 1 Requests Per Second > Higher Is Better A . 6546.58 |========================================================= B . 7614.76 |================================================================== Apache HTTP Server 2.4.48 Concurrent Requests: 20 Requests Per Second > Higher Is Better A . 16782.58 |============================================================ B . 18160.31 |================================================================= Apache HTTP Server 2.4.48 Concurrent Requests: 100 Requests Per Second > Higher Is Better A . 37378.82 |================================================================= B . 36149.79 |=============================================================== Apache HTTP Server 2.4.48 Concurrent Requests: 200 Requests Per Second > Higher Is Better A . 45403.03 |========================================================== B . 50848.21 |================================================================= Apache HTTP Server 2.4.48 Concurrent Requests: 500 Requests Per Second > Higher Is Better A . 51287.26 |================================================================ B . 52228.94 |================================================================= Apache HTTP Server 2.4.48 Concurrent Requests: 1000 Requests Per Second > Higher Is Better A . 44329.50 |============================================================ B . 48169.54 |================================================================= InfluxDB 1.8.2 Concurrent Streams: 4 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 10000 val/sec > Higher Is Better InfluxDB 1.8.2 Concurrent Streams: 64 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 10000 val/sec > Higher Is Better InfluxDB 1.8.2 Concurrent Streams: 1024 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 10000 val/sec > Higher Is Better