christmas comet Intel Core i7-10700T testing with a Logic Supply RXM-181 (Z01-0002A026 BIOS) and Intel UHD 630 CML GT2 30GB on Ubuntu 22.04 via the Phoronix Test Suite.
HTML result view exported from: https://openbenchmarking.org/result/2212231-NE-CHRISTMAS95&grr&rdt .
christmas comet Processor Motherboard Chipset Memory Disk Graphics Audio Monitor Network OS Kernel Desktop Display Server OpenGL OpenCL Vulkan Compiler File-System Screen Resolution a b c Intel Core i7-10700T @ 4.50GHz (8 Cores / 16 Threads) Logic Supply RXM-181 (Z01-0002A026 BIOS) Intel Comet Lake PCH 32GB 256GB TS256GMTS800 Intel UHD 630 CML GT2 30GB (1200MHz) Realtek ALC233 DELL P2415Q Intel I219-LM + Intel I210 Ubuntu 22.04 5.15.0-52-generic (x86_64) GNOME Shell 42.2 X Server + Wayland 4.6 Mesa 22.0.1 OpenCL 3.0 1.3.204 GCC 11.3.0 ext4 1920x1080 OpenBenchmarking.org Kernel Details - Transparent Huge Pages: madvise Compiler Details - --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-xKiWfi/gcc-11-11.3.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-xKiWfi/gcc-11-11.3.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details - Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0xf0 - Thermald 2.4.9 Python Details - Python 3.10.6 Security Details - itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Mitigation of Clear buffers; SMT vulnerable + retbleed: Mitigation of Enhanced IBRS + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Mitigation of Microcode + tsx_async_abort: Not affected
christmas comet scikit-learn: Sparse Rand Projections, 100 Iterations blender: Barbershop - CPU-Only build-linux-kernel: allmodconfig blender: Pabellon Barcelona - CPU-Only nekrs: TurboPipe Periodic fluidx3d: FP32-FP16C blender: Classroom - CPU-Only fluidx3d: FP32-FP32 fluidx3d: FP32-FP16S blender: Fishy Cat - CPU-Only openvkl: vklBenchmark ISPC openvkl: vklBenchmark Scalar numenta-nab: KNN CAD blender: BMW27 - CPU-Only scikit-learn: MNIST Dataset stargate: 192000 - 512 numenta-nab: Earthgecko Skyline stargate: 192000 - 1024 build-linux-kernel: defconfig stargate: 96000 - 512 svt-av1: Preset 4 - Bosphorus 4K stargate: 96000 - 1024 scikit-learn: TSNE MNIST Dataset cockroach: KV, 10% Reads - 1024 cockroach: KV, 60% Reads - 1024 cockroach: KV, 50% Reads - 1024 cockroach: KV, 95% Reads - 1024 cockroach: KV, 10% Reads - 512 cockroach: KV, 60% Reads - 512 cockroach: KV, 50% Reads - 512 cockroach: KV, 95% Reads - 512 cockroach: KV, 50% Reads - 256 cockroach: KV, 95% Reads - 256 cockroach: KV, 10% Reads - 256 cockroach: KV, 60% Reads - 256 cockroach: KV, 50% Reads - 128 cockroach: KV, 60% Reads - 128 cockroach: KV, 95% Reads - 128 cockroach: KV, 10% Reads - 128 onednn: Recurrent Neural Network Training - f32 - CPU onednn: Recurrent Neural Network Training - u8s8f32 - CPU onednn: Recurrent Neural Network Training - bf16bf16bf16 - CPU cockroach: MoVR - 128 cockroach: MoVR - 256 cockroach: MoVR - 1024 cockroach: MoVR - 512 stargate: 480000 - 512 stargate: 44100 - 512 stargate: 480000 - 1024 stargate: 44100 - 1024 onednn: Recurrent Neural Network Inference - u8s8f32 - CPU onednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPU onednn: Recurrent Neural Network Inference - f32 - CPU numenta-nab: Contextual Anomaly Detector OSE openvino: Person Detection FP32 - CPU openvino: Person Detection FP32 - CPU openvino: Person Detection FP16 - CPU openvino: Person Detection FP16 - CPU openvino: Face Detection FP16 - CPU openvino: Face Detection FP16 - CPU openvino: Face Detection FP16-INT8 - CPU openvino: Face Detection FP16-INT8 - CPU openvino: Machine Translation EN To DE FP16 - CPU openvino: Machine Translation EN To DE FP16 - CPU openvino: Person Vehicle Bike Detection FP16 - CPU openvino: Person Vehicle Bike Detection FP16 - CPU openvino: Vehicle Detection FP16-INT8 - CPU openvino: Vehicle Detection FP16-INT8 - CPU openvino: Vehicle Detection FP16 - CPU openvino: Vehicle Detection FP16 - CPU openvino: Weld Porosity Detection FP16 - CPU openvino: Weld Porosity Detection FP16 - CPU openvino: Weld Porosity Detection FP16-INT8 - CPU openvino: Weld Porosity Detection FP16-INT8 - CPU openvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPU openvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPU openvino: Age Gender Recognition Retail 0013 FP16 - CPU openvino: Age Gender Recognition Retail 0013 FP16 - CPU numenta-nab: Bayesian Changepoint rav1e: 1 svt-av1: Preset 8 - Bosphorus 4K svt-av1: Preset 4 - Bosphorus 1080p numenta-nab: Relative Entropy rav1e: 5 onednn: Deconvolution Batch shapes_1d - f32 - CPU onednn: Deconvolution Batch shapes_1d - u8s8f32 - CPU rav1e: 6 numenta-nab: Windowed Gaussian onednn: IP Shapes 1D - f32 - CPU onednn: IP Shapes 1D - u8s8f32 - CPU onednn: Matrix Multiply Batch Shapes Transformer - f32 - CPU onednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU svt-av1: Preset 8 - Bosphorus 1080p rav1e: 10 onednn: IP Shapes 3D - f32 - CPU onednn: IP Shapes 3D - u8s8f32 - CPU svt-av1: Preset 12 - Bosphorus 4K svt-av1: Preset 13 - Bosphorus 4K onednn: Convolution Batch Shapes Auto - f32 - CPU onednn: Convolution Batch Shapes Auto - u8s8f32 - CPU onednn: Deconvolution Batch shapes_3d - f32 - CPU onednn: Deconvolution Batch shapes_3d - u8s8f32 - CPU svt-av1: Preset 12 - Bosphorus 1080p svt-av1: Preset 13 - Bosphorus 1080p onednn: IP Shapes 1D - bf16bf16bf16 - CPU a b c 3294.519 3260.14 2642.392 1083.6 25130200000 173 902.53 201 357 417.83 93 46 364.397 306.09 251.412 0.791489 200.783 0.867297 190.421 1.204979 1.229 1.293792 109.345 21256.4 24536.7 23641.5 28002.8 22167.5 25923.7 25227.2 29621.9 25273.3 30543 21256.3 26268.4 21728.4 24369.6 31139.2 12318.5 6850.49 6878.39 6869.52 174.4 174.8 173.6 173.9 1.638868 1.69939 1.731054 1.798704 3622.8 3640.28 3625.06 69.885 4627.4 0.85 4564.3 0.86 2907.95 1.37 1552.62 2.57 268.31 14.9 22.07 181.05 24.25 164.79 37.9 105.44 32.08 124.59 30.73 260.1 2.03 3914.1 2.2 3609.48 56.706 0.413 16.499 4.392 32.183 2.238 11.6929 3.80985 3.125 18.272 5.11104 2.37749 3.99763 2.23303 51.631 8.33 10.8363 2.48501 73.506 79.105 17.1397 15.5046 8.30947 5.02536 323.225 356.085 3294.87 3257.91 2642.66 1086.78 25286500000 175 904.16 201 377 417.45 94 46 364.876 305.77 252.299 0.794756 193.302 0.870159 189.627 1.206529 1.247 1.29592 108.375 21125.9 24424.9 23487.2 27948 22053.8 25936.5 25164.2 29633.5 25267.5 30259.7 21051.4 26246.9 21711.8 24360.3 30808.9 12364.9 6861.12 6904.88 6897.51 175.2 175.2 171.6 172.1 1.651464 1.717714 1.733011 1.802469 3783.41 3657.32 3632.36 70.238 4705.67 0.85 4567.98 0.86 2900.79 1.37 1555.24 2.57 267.76 14.92 22.52 177.44 24.35 164.14 38.33 104.28 31.98 124.97 30.96 258.24 2.03 3925.52 2.21 3595.44 60.654 0.417 16.693 4.433 32.448 2.269 12.5886 3.77816 3.163 18.344 5.03425 2.26968 4.01087 2.2057 53.451 8.621 10.8813 2.47157 75.452 80.31 17.1202 15.5676 8.38301 5.05109 324.322 356.327 3302.545 3234.94 2688.942 1083.92 25386800000 175 905.71 200 381 417.82 94 46 360.585 307.76 251.815 0.792159 201.311 0.870579 189.634 1.206909 1.247 1.295966 110.002 21160.8 24231 23596 27834.5 22080.3 26058.7 25045.3 29572.8 25133.1 30253.2 21050.4 26188 21625.8 24267.8 30644.9 12422.9 6832.22 6850.47 6842.96 173.1 167 168.9 173.1 1.649194 1.704004 1.737779 1.802314 3612.19 3636.49 3629.54 69.625 4632.94 0.85 4555.58 0.86 2889.56 1.38 1548.9 2.58 267.65 14.93 22.85 174.87 24.35 164.12 38.12 104.85 31.85 125.45 31.03 257.58 2 3970.02 2.2 3620.42 60.22 0.416 16.673 4.446 34.335 2.284 11.7109 3.78238 3.155 18.648 67.4107 40.0963 3.99005 2.14039 53.145 8.654 47.6781 51.5073 74.668 80.973 17.2961 15.4711 8.47927 5.15364 328.002 364.038 OpenBenchmarking.org
Scikit-Learn Benchmark: Sparse Random Projections, 100 Iterations OpenBenchmarking.org Seconds, Fewer Is Better Scikit-Learn 1.1.3 Benchmark: Sparse Random Projections, 100 Iterations a b c 700 1400 2100 2800 3500 3294.52 3294.87 3302.55
Blender Blend File: Barbershop - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 3.4 Blend File: Barbershop - Compute: CPU-Only a b c 700 1400 2100 2800 3500 3260.14 3257.91 3234.94
Timed Linux Kernel Compilation Build: allmodconfig OpenBenchmarking.org Seconds, Fewer Is Better Timed Linux Kernel Compilation 6.1 Build: allmodconfig a b c 600 1200 1800 2400 3000 2642.39 2642.66 2688.94
Blender Blend File: Pabellon Barcelona - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 3.4 Blend File: Pabellon Barcelona - Compute: CPU-Only a b c 200 400 600 800 1000 1083.60 1086.78 1083.92
nekRS Input: TurboPipe Periodic OpenBenchmarking.org FLOP/s, More Is Better nekRS 22.0 Input: TurboPipe Periodic a b c 5000M 10000M 15000M 20000M 25000M 25130200000 25286500000 25386800000 1. (CXX) g++ options: -fopenmp -O2 -march=native -mtune=native -ftree-vectorize -lmpi_cxx -lmpi
FluidX3D Test: FP32-FP16C OpenBenchmarking.org MLUPs/s, More Is Better FluidX3D 1.4 Test: FP32-FP16C a b c 40 80 120 160 200 173 175 175
Blender Blend File: Classroom - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 3.4 Blend File: Classroom - Compute: CPU-Only a b c 200 400 600 800 1000 902.53 904.16 905.71
FluidX3D Test: FP32-FP32 OpenBenchmarking.org MLUPs/s, More Is Better FluidX3D 1.4 Test: FP32-FP32 a b c 40 80 120 160 200 201 201 200
FluidX3D Test: FP32-FP16S OpenBenchmarking.org MLUPs/s, More Is Better FluidX3D 1.4 Test: FP32-FP16S a b c 80 160 240 320 400 357 377 381
Blender Blend File: Fishy Cat - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 3.4 Blend File: Fishy Cat - Compute: CPU-Only a b c 90 180 270 360 450 417.83 417.45 417.82
OpenVKL Benchmark: vklBenchmark ISPC OpenBenchmarking.org Items / Sec, More Is Better OpenVKL 1.3.1 Benchmark: vklBenchmark ISPC a b c 20 40 60 80 100 93 94 94 MIN: 10 / MAX: 1436 MIN: 10 / MAX: 1446 MIN: 10 / MAX: 1453
OpenVKL Benchmark: vklBenchmark Scalar OpenBenchmarking.org Items / Sec, More Is Better OpenVKL 1.3.1 Benchmark: vklBenchmark Scalar a b c 10 20 30 40 50 46 46 46 MIN: 5 / MAX: 1052 MIN: 5 / MAX: 1072 MIN: 5 / MAX: 1066
Numenta Anomaly Benchmark Detector: KNN CAD OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: KNN CAD a b c 80 160 240 320 400 364.40 364.88 360.59
Blender Blend File: BMW27 - Compute: CPU-Only OpenBenchmarking.org Seconds, Fewer Is Better Blender 3.4 Blend File: BMW27 - Compute: CPU-Only a b c 70 140 210 280 350 306.09 305.77 307.76
Scikit-Learn Benchmark: MNIST Dataset OpenBenchmarking.org Seconds, Fewer Is Better Scikit-Learn 1.1.3 Benchmark: MNIST Dataset a b c 60 120 180 240 300 251.41 252.30 251.82
Stargate Digital Audio Workstation Sample Rate: 192000 - Buffer Size: 512 OpenBenchmarking.org Render Ratio, More Is Better Stargate Digital Audio Workstation 22.11.5 Sample Rate: 192000 - Buffer Size: 512 a b c 0.1788 0.3576 0.5364 0.7152 0.894 0.791489 0.794756 0.792159 1. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions
Numenta Anomaly Benchmark Detector: Earthgecko Skyline OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: Earthgecko Skyline a b c 40 80 120 160 200 200.78 193.30 201.31
Stargate Digital Audio Workstation Sample Rate: 192000 - Buffer Size: 1024 OpenBenchmarking.org Render Ratio, More Is Better Stargate Digital Audio Workstation 22.11.5 Sample Rate: 192000 - Buffer Size: 1024 a b c 0.1959 0.3918 0.5877 0.7836 0.9795 0.867297 0.870159 0.870579 1. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions
Timed Linux Kernel Compilation Build: defconfig OpenBenchmarking.org Seconds, Fewer Is Better Timed Linux Kernel Compilation 6.1 Build: defconfig a b c 40 80 120 160 200 190.42 189.63 189.63
Stargate Digital Audio Workstation Sample Rate: 96000 - Buffer Size: 512 OpenBenchmarking.org Render Ratio, More Is Better Stargate Digital Audio Workstation 22.11.5 Sample Rate: 96000 - Buffer Size: 512 a b c 0.2716 0.5432 0.8148 1.0864 1.358 1.204979 1.206529 1.206909 1. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions
SVT-AV1 Encoder Mode: Preset 4 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.4 Encoder Mode: Preset 4 - Input: Bosphorus 4K a b c 0.2806 0.5612 0.8418 1.1224 1.403 1.229 1.247 1.247 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
Stargate Digital Audio Workstation Sample Rate: 96000 - Buffer Size: 1024 OpenBenchmarking.org Render Ratio, More Is Better Stargate Digital Audio Workstation 22.11.5 Sample Rate: 96000 - Buffer Size: 1024 a b c 0.2916 0.5832 0.8748 1.1664 1.458 1.293792 1.295920 1.295966 1. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions
Scikit-Learn Benchmark: TSNE MNIST Dataset OpenBenchmarking.org Seconds, Fewer Is Better Scikit-Learn 1.1.3 Benchmark: TSNE MNIST Dataset a b c 20 40 60 80 100 109.35 108.38 110.00
CockroachDB Workload: KV, 10% Reads - Concurrency: 1024 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 10% Reads - Concurrency: 1024 a b c 5K 10K 15K 20K 25K 21256.4 21125.9 21160.8
CockroachDB Workload: KV, 60% Reads - Concurrency: 1024 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 60% Reads - Concurrency: 1024 a b c 5K 10K 15K 20K 25K 24536.7 24424.9 24231.0
CockroachDB Workload: KV, 50% Reads - Concurrency: 1024 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 50% Reads - Concurrency: 1024 a b c 5K 10K 15K 20K 25K 23641.5 23487.2 23596.0
CockroachDB Workload: KV, 95% Reads - Concurrency: 1024 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 95% Reads - Concurrency: 1024 a b c 6K 12K 18K 24K 30K 28002.8 27948.0 27834.5
CockroachDB Workload: KV, 10% Reads - Concurrency: 512 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 10% Reads - Concurrency: 512 a b c 5K 10K 15K 20K 25K 22167.5 22053.8 22080.3
CockroachDB Workload: KV, 60% Reads - Concurrency: 512 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 60% Reads - Concurrency: 512 a b c 6K 12K 18K 24K 30K 25923.7 25936.5 26058.7
CockroachDB Workload: KV, 50% Reads - Concurrency: 512 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 50% Reads - Concurrency: 512 a b c 5K 10K 15K 20K 25K 25227.2 25164.2 25045.3
CockroachDB Workload: KV, 95% Reads - Concurrency: 512 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 95% Reads - Concurrency: 512 a b c 6K 12K 18K 24K 30K 29621.9 29633.5 29572.8
CockroachDB Workload: KV, 50% Reads - Concurrency: 256 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 50% Reads - Concurrency: 256 a b c 5K 10K 15K 20K 25K 25273.3 25267.5 25133.1
CockroachDB Workload: KV, 95% Reads - Concurrency: 256 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 95% Reads - Concurrency: 256 a b c 7K 14K 21K 28K 35K 30543.0 30259.7 30253.2
CockroachDB Workload: KV, 10% Reads - Concurrency: 256 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 10% Reads - Concurrency: 256 a b c 5K 10K 15K 20K 25K 21256.3 21051.4 21050.4
CockroachDB Workload: KV, 60% Reads - Concurrency: 256 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 60% Reads - Concurrency: 256 a b c 6K 12K 18K 24K 30K 26268.4 26246.9 26188.0
CockroachDB Workload: KV, 50% Reads - Concurrency: 128 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 50% Reads - Concurrency: 128 a b c 5K 10K 15K 20K 25K 21728.4 21711.8 21625.8
CockroachDB Workload: KV, 60% Reads - Concurrency: 128 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 60% Reads - Concurrency: 128 a b c 5K 10K 15K 20K 25K 24369.6 24360.3 24267.8
CockroachDB Workload: KV, 95% Reads - Concurrency: 128 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 95% Reads - Concurrency: 128 a b c 7K 14K 21K 28K 35K 31139.2 30808.9 30644.9
CockroachDB Workload: KV, 10% Reads - Concurrency: 128 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: KV, 10% Reads - Concurrency: 128 a b c 3K 6K 9K 12K 15K 12318.5 12364.9 12422.9
oneDNN Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU a b c 1500 3000 4500 6000 7500 6850.49 6861.12 6832.22 MIN: 6698.29 MIN: 6709.14 MIN: 6683.82 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU a b c 1500 3000 4500 6000 7500 6878.39 6904.88 6850.47 MIN: 6697.57 MIN: 6753.33 MIN: 6689.16 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU a b c 1500 3000 4500 6000 7500 6869.52 6897.51 6842.96 MIN: 6707.44 MIN: 6737.21 MIN: 6686.11 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
CockroachDB Workload: MoVR - Concurrency: 128 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: MoVR - Concurrency: 128 a b c 40 80 120 160 200 174.4 175.2 173.1
CockroachDB Workload: MoVR - Concurrency: 256 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: MoVR - Concurrency: 256 a b c 40 80 120 160 200 174.8 175.2 167.0
CockroachDB Workload: MoVR - Concurrency: 1024 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: MoVR - Concurrency: 1024 a b c 40 80 120 160 200 173.6 171.6 168.9
CockroachDB Workload: MoVR - Concurrency: 512 OpenBenchmarking.org ops/s, More Is Better CockroachDB 22.2 Workload: MoVR - Concurrency: 512 a b c 40 80 120 160 200 173.9 172.1 173.1
Stargate Digital Audio Workstation Sample Rate: 480000 - Buffer Size: 512 OpenBenchmarking.org Render Ratio, More Is Better Stargate Digital Audio Workstation 22.11.5 Sample Rate: 480000 - Buffer Size: 512 a b c 0.3716 0.7432 1.1148 1.4864 1.858 1.638868 1.651464 1.649194 1. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions
Stargate Digital Audio Workstation Sample Rate: 44100 - Buffer Size: 512 OpenBenchmarking.org Render Ratio, More Is Better Stargate Digital Audio Workstation 22.11.5 Sample Rate: 44100 - Buffer Size: 512 a b c 0.3865 0.773 1.1595 1.546 1.9325 1.699390 1.717714 1.704004 1. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions
Stargate Digital Audio Workstation Sample Rate: 480000 - Buffer Size: 1024 OpenBenchmarking.org Render Ratio, More Is Better Stargate Digital Audio Workstation 22.11.5 Sample Rate: 480000 - Buffer Size: 1024 a b c 0.391 0.782 1.173 1.564 1.955 1.731054 1.733011 1.737779 1. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions
Stargate Digital Audio Workstation Sample Rate: 44100 - Buffer Size: 1024 OpenBenchmarking.org Render Ratio, More Is Better Stargate Digital Audio Workstation 22.11.5 Sample Rate: 44100 - Buffer Size: 1024 a b c 0.4056 0.8112 1.2168 1.6224 2.028 1.798704 1.802469 1.802314 1. (CXX) g++ options: -lpthread -lsndfile -lm -O3 -march=native -ffast-math -funroll-loops -fstrength-reduce -fstrict-aliasing -finline-functions
oneDNN Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU a b c 800 1600 2400 3200 4000 3622.80 3783.41 3612.19 MIN: 3501.52 MIN: 3514.1 MIN: 3493.86 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU a b c 800 1600 2400 3200 4000 3640.28 3657.32 3636.49 MIN: 3508.09 MIN: 3527.16 MIN: 3511.21 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU a b c 800 1600 2400 3200 4000 3625.06 3632.36 3629.54 MIN: 3495.85 MIN: 3504.4 MIN: 3500.37 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
Numenta Anomaly Benchmark Detector: Contextual Anomaly Detector OSE OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: Contextual Anomaly Detector OSE a b c 16 32 48 64 80 69.89 70.24 69.63
OpenVINO Model: Person Detection FP32 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2022.3 Model: Person Detection FP32 - Device: CPU a b c 1000 2000 3000 4000 5000 4627.40 4705.67 4632.94 MIN: 3461.83 / MAX: 5004.23 MIN: 3414.61 / MAX: 4954.48 MIN: 3246.77 / MAX: 4955.57 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Person Detection FP32 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2022.3 Model: Person Detection FP32 - Device: CPU a b c 0.1913 0.3826 0.5739 0.7652 0.9565 0.85 0.85 0.85 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Person Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2022.3 Model: Person Detection FP16 - Device: CPU a b c 1000 2000 3000 4000 5000 4564.30 4567.98 4555.58 MIN: 3395.41 / MAX: 4924.67 MIN: 3311.82 / MAX: 4914.52 MIN: 3391.76 / MAX: 4917.08 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Person Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2022.3 Model: Person Detection FP16 - Device: CPU a b c 0.1935 0.387 0.5805 0.774 0.9675 0.86 0.86 0.86 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Face Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2022.3 Model: Face Detection FP16 - Device: CPU a b c 600 1200 1800 2400 3000 2907.95 2900.79 2889.56 MIN: 2048.8 / MAX: 3076.74 MIN: 1935.99 / MAX: 3085.36 MIN: 2130.02 / MAX: 3082.08 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Face Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2022.3 Model: Face Detection FP16 - Device: CPU a b c 0.3105 0.621 0.9315 1.242 1.5525 1.37 1.37 1.38 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Face Detection FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2022.3 Model: Face Detection FP16-INT8 - Device: CPU a b c 300 600 900 1200 1500 1552.62 1555.24 1548.90 MIN: 994.18 / MAX: 1658.73 MIN: 987.74 / MAX: 1638.42 MIN: 1000.5 / MAX: 1643.06 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Face Detection FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2022.3 Model: Face Detection FP16-INT8 - Device: CPU a b c 0.5805 1.161 1.7415 2.322 2.9025 2.57 2.57 2.58 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Machine Translation EN To DE FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2022.3 Model: Machine Translation EN To DE FP16 - Device: CPU a b c 60 120 180 240 300 268.31 267.76 267.65 MIN: 173.59 / MAX: 305.62 MIN: 157 / MAX: 354.66 MIN: 158.54 / MAX: 333.85 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Machine Translation EN To DE FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2022.3 Model: Machine Translation EN To DE FP16 - Device: CPU a b c 4 8 12 16 20 14.90 14.92 14.93 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Person Vehicle Bike Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2022.3 Model: Person Vehicle Bike Detection FP16 - Device: CPU a b c 5 10 15 20 25 22.07 22.52 22.85 MIN: 12.92 / MAX: 53.87 MIN: 12.96 / MAX: 56.58 MIN: 13.11 / MAX: 56.73 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Person Vehicle Bike Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2022.3 Model: Person Vehicle Bike Detection FP16 - Device: CPU a b c 40 80 120 160 200 181.05 177.44 174.87 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Vehicle Detection FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2022.3 Model: Vehicle Detection FP16-INT8 - Device: CPU a b c 6 12 18 24 30 24.25 24.35 24.35 MIN: 14.38 / MAX: 63.86 MIN: 15.05 / MAX: 64.68 MIN: 14.89 / MAX: 63.77 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Vehicle Detection FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2022.3 Model: Vehicle Detection FP16-INT8 - Device: CPU a b c 40 80 120 160 200 164.79 164.14 164.12 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Vehicle Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2022.3 Model: Vehicle Detection FP16 - Device: CPU a b c 9 18 27 36 45 37.90 38.33 38.12 MIN: 25.27 / MAX: 84.62 MIN: 26.97 / MAX: 86.94 MIN: 26.05 / MAX: 83.42 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Vehicle Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2022.3 Model: Vehicle Detection FP16 - Device: CPU a b c 20 40 60 80 100 105.44 104.28 104.85 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Weld Porosity Detection FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2022.3 Model: Weld Porosity Detection FP16 - Device: CPU a b c 7 14 21 28 35 32.08 31.98 31.85 MIN: 19.86 / MAX: 78.3 MIN: 18.51 / MAX: 81.22 MIN: 20.23 / MAX: 81.16 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Weld Porosity Detection FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2022.3 Model: Weld Porosity Detection FP16 - Device: CPU a b c 30 60 90 120 150 124.59 124.97 125.45 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Weld Porosity Detection FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2022.3 Model: Weld Porosity Detection FP16-INT8 - Device: CPU a b c 7 14 21 28 35 30.73 30.96 31.03 MIN: 16.95 / MAX: 84.01 MIN: 18.67 / MAX: 47.56 MIN: 12.41 / MAX: 84.78 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Weld Porosity Detection FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2022.3 Model: Weld Porosity Detection FP16-INT8 - Device: CPU a b c 60 120 180 240 300 260.10 258.24 257.58 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2022.3 Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU a b c 0.4568 0.9136 1.3704 1.8272 2.284 2.03 2.03 2.00 MIN: 0.7 / MAX: 26.99 MIN: 0.7 / MAX: 5.69 MIN: 0.66 / MAX: 19.69 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2022.3 Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU a b c 900 1800 2700 3600 4500 3914.10 3925.52 3970.02 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU OpenBenchmarking.org ms, Fewer Is Better OpenVINO 2022.3 Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU a b c 0.4973 0.9946 1.4919 1.9892 2.4865 2.20 2.21 2.20 MIN: 0.69 / MAX: 20 MIN: 0.73 / MAX: 19.99 MIN: 0.75 / MAX: 7.41 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
OpenVINO Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU OpenBenchmarking.org FPS, More Is Better OpenVINO 2022.3 Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU a b c 800 1600 2400 3200 4000 3609.48 3595.44 3620.42 1. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared
Numenta Anomaly Benchmark Detector: Bayesian Changepoint OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: Bayesian Changepoint a b c 14 28 42 56 70 56.71 60.65 60.22
rav1e Speed: 1 OpenBenchmarking.org Frames Per Second, More Is Better rav1e 0.6.1 Speed: 1 a b c 0.0938 0.1876 0.2814 0.3752 0.469 0.413 0.417 0.416
SVT-AV1 Encoder Mode: Preset 8 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.4 Encoder Mode: Preset 8 - Input: Bosphorus 4K a b c 4 8 12 16 20 16.50 16.69 16.67 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 4 - Input: Bosphorus 1080p OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.4 Encoder Mode: Preset 4 - Input: Bosphorus 1080p a b c 1.0004 2.0008 3.0012 4.0016 5.002 4.392 4.433 4.446 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
Numenta Anomaly Benchmark Detector: Relative Entropy OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: Relative Entropy a b c 8 16 24 32 40 32.18 32.45 34.34
rav1e Speed: 5 OpenBenchmarking.org Frames Per Second, More Is Better rav1e 0.6.1 Speed: 5 a b c 0.5139 1.0278 1.5417 2.0556 2.5695 2.238 2.269 2.284
oneDNN Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU a b c 3 6 9 12 15 11.69 12.59 11.71 MIN: 7.4 MIN: 7.47 MIN: 7.33 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU a b c 0.8572 1.7144 2.5716 3.4288 4.286 3.80985 3.77816 3.78238 MIN: 2.87 MIN: 2.87 MIN: 2.9 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
rav1e Speed: 6 OpenBenchmarking.org Frames Per Second, More Is Better rav1e 0.6.1 Speed: 6 a b c 0.7117 1.4234 2.1351 2.8468 3.5585 3.125 3.163 3.155
Numenta Anomaly Benchmark Detector: Windowed Gaussian OpenBenchmarking.org Seconds, Fewer Is Better Numenta Anomaly Benchmark 1.1 Detector: Windowed Gaussian a b c 5 10 15 20 25 18.27 18.34 18.65
oneDNN Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU a b c 15 30 45 60 75 5.11104 5.03425 67.41070 MIN: 4.43 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU a b c 9 18 27 36 45 2.37749 2.26968 40.09630 MIN: 1.93 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPU a b c 0.9024 1.8048 2.7072 3.6096 4.512 3.99763 4.01087 3.99005 MIN: 3.84 MIN: 3.89 MIN: 3.8 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU a b c 0.5024 1.0048 1.5072 2.0096 2.512 2.23303 2.20570 2.14039 MIN: 1.66 MIN: 1.65 MIN: 1.66 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
SVT-AV1 Encoder Mode: Preset 8 - Input: Bosphorus 1080p OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.4 Encoder Mode: Preset 8 - Input: Bosphorus 1080p a b c 12 24 36 48 60 51.63 53.45 53.15 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
rav1e Speed: 10 OpenBenchmarking.org Frames Per Second, More Is Better rav1e 0.6.1 Speed: 10 a b c 2 4 6 8 10 8.330 8.621 8.654
oneDNN Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU a b c 11 22 33 44 55 10.84 10.88 47.68 MIN: 10.2 MIN: 10.22 MIN: 10.77 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU a b c 12 24 36 48 60 2.48501 2.47157 51.50730 MIN: 2.51 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
SVT-AV1 Encoder Mode: Preset 12 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.4 Encoder Mode: Preset 12 - Input: Bosphorus 4K a b c 20 40 60 80 100 73.51 75.45 74.67 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 13 - Input: Bosphorus 4K OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.4 Encoder Mode: Preset 13 - Input: Bosphorus 4K a b c 20 40 60 80 100 79.11 80.31 80.97 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
oneDNN Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU a b c 4 8 12 16 20 17.14 17.12 17.30 MIN: 17.01 MIN: 16.99 MIN: 17 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU a b c 4 8 12 16 20 15.50 15.57 15.47 MIN: 14.98 MIN: 15.25 MIN: 15.23 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU a b c 2 4 6 8 10 8.30947 8.38301 8.47927 MIN: 7.97 MIN: 8.01 MIN: 8.09 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
oneDNN Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU OpenBenchmarking.org ms, Fewer Is Better oneDNN 3.0 Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU a b c 1.1596 2.3192 3.4788 4.6384 5.798 5.02536 5.05109 5.15364 MIN: 4.65 MIN: 4.66 MIN: 4.63 1. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl
SVT-AV1 Encoder Mode: Preset 12 - Input: Bosphorus 1080p OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.4 Encoder Mode: Preset 12 - Input: Bosphorus 1080p a b c 70 140 210 280 350 323.23 324.32 328.00 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
SVT-AV1 Encoder Mode: Preset 13 - Input: Bosphorus 1080p OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 1.4 Encoder Mode: Preset 13 - Input: Bosphorus 1080p a b c 80 160 240 320 400 356.09 356.33 364.04 1. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq
Phoronix Test Suite v10.8.5