core-i5-12400-april

Intel Core i5-12400 testing with a ASRock B660M-HDV (3.02 BIOS) and llvmpipe on Ubuntu 22.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2204077-NE-COREI512423
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results
Show Result Confidence Charts
Allow Limiting Results To Certain Suite(s)

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Disable Color Branding
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Toggle/Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
A
April 06 2022
  1 Hour, 59 Minutes
Intel Core i5-12400
April 06 2022
  6 Minutes
B
April 06 2022
  5 Hours, 40 Minutes
C
April 07 2022
  5 Hours, 38 Minutes
Invert Behavior (Only Show Selected Data)
  3 Hours, 21 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


core-i5-12400-april ProcessorMotherboardChipsetMemoryDiskGraphicsAudioNetworkOSKernelDesktopDisplay ServerOpenGLVulkanCompilerFile-SystemScreen ResolutionAIntel Core i5-12400BCIntel Core i5-12400 @ 5.60GHz (6 Cores / 12 Threads)ASRock B660M-HDV (3.02 BIOS)Intel Device 7aa716GB512GB SabrentllvmpipeRealtek ALC897IntelUbuntu 22.045.15.0-18-generic (x86_64)GNOME Shell 41.3X Server 1.20.144.5 Mesa 21.2.2 (LLVM 12.0.1 256 bits)1.1.182GCC 11.2.0ext41920x10804.5 Mesa 22.0.1 (LLVM 13.0.1 256 bits)1.2.204OpenBenchmarking.orgKernel Details- Transparent Huge Pages: madviseCompiler Details- A: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-iOLsLC/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-iOLsLC/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Intel Core i5-12400: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - B: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - C: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- A: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.7- Intel Core i5-12400: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.9- B: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.9- C: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.9Java Details- A, B, C: OpenJDK Runtime Environment (build 11.0.14.1+1-Ubuntu-0ubuntu1)Python Details- A: Python 3.9.12- B: Python 3.10.4- C: Python 3.10.4Security Details- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected

core-i5-12400-april onednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: IP Shapes 3D - f32 - CPUonednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUonednn: IP Shapes 1D - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUonnx: fcn-resnet101-11 - CPU - Parallelaom-av1: Speed 10 Realtime - Bosphorus 4Konnx: super-resolution-10 - CPU - Parallelaom-av1: Speed 10 Realtime - Bosphorus 1080pospray: particle_volume/scivis/real_timeaom-av1: Speed 4 Two-Pass - Bosphorus 4Kospray-studio: 2 - 1080p - 16 - Path Traceraom-av1: Speed 6 Two-Pass - Bosphorus 4Kospray-studio: 1 - 1080p - 16 - Path Tracerospray-studio: 1 - 1080p - 32 - Path Tracerdav1d: Summer Nature 1080pospray-studio: 3 - 1080p - 1 - Path Tracerospray-studio: 2 - 1080p - 32 - Path Tracerospray-studio: 3 - 1080p - 32 - Path Traceraom-av1: Speed 6 Two-Pass - Bosphorus 1080pdav1d: Chimera 1080p 10-bitospray: particle_volume/pathtracer/real_timedav1d: Summer Nature 4Kcompress-pbzip2: FreeBSD-13.0-RELEASE-amd64-memstick.img Compressiononnx: ArcFace ResNet-100 - CPU - Parallelospray: gravity_spheres_volume/dim_512/pathtracer/real_timeospray-studio: 3 - 1080p - 16 - Path Tracerbuild-wasmer: Time To Compileospray-studio: 2 - 1080p - 1 - Path Tracerjava-jmh: Throughputospray: particle_volume/ao/real_timeospray-studio: 1 - 1080p - 1 - Path Tracerrocksdb: Update Randospray: gravity_spheres_volume/dim_512/scivis/real_timedav1d: Chimera 1080procksdb: Read While Writingrocksdb: Read Rand Write Randospray: gravity_spheres_volume/dim_512/ao/real_timeaom-av1: Speed 8 Realtime - Bosphorus 1080paom-av1: Speed 0 Two-Pass - Bosphorus 1080ponnx: yolov4 - CPU - Parallelbuild-mplayer: Time To Compilerocksdb: Rand Readaom-av1: Speed 4 Two-Pass - Bosphorus 1080paom-av1: Speed 6 Realtime - Bosphorus 4Kaom-av1: Speed 9 Realtime - Bosphorus 4Kaom-av1: Speed 0 Two-Pass - Bosphorus 4Kaom-av1: Speed 8 Realtime - Bosphorus 4Kaom-av1: Speed 6 Realtime - Bosphorus 1080ponnx: yolov4 - CPU - Standardonnx: GPT-2 - CPU - Parallelonnx: super-resolution-10 - CPU - Standardonnx: fcn-resnet101-11 - CPU - Standardonnx: GPT-2 - CPU - Standardonnx: bertsquad-12 - CPU - Standardonnx: bertsquad-12 - CPU - Parallelonnx: ArcFace ResNet-100 - CPU - Standardavifenc: 2avifenc: 10, Losslessavifenc: 0avifenc: 6, Losslessavifenc: 6onednn: Deconvolution Batch shapes_1d - f32 - CPUaom-av1: Speed 9 Realtime - Bosphorus 1080pAIntel Core i5-12400BC36805.217.215234757.429463.364.373746.670153.701146.3983113.72450.702447.693728.326871.826834134.527.917228488.224467.54461.382845133.2112.32034.1757299856352115145710.02418911668913692924.11481.75163.591172.7110.9558532.472226856885.485347120972409216.50212.968833986797801.79913561.82146716614975541.83804107.060.3427742.432456972069.5813.9163.160.1342.4711.6729257452822426878439427869162.853124.685.166.2189.7414.91812.2074099.421.981564102.344095.969.901067.698901.047161.426413.7405213.756514.33180.9578032.689662264.902.108942264.982271.695672.383247150.3313.88934.70508978.9950151102548795.53374010412412236626.97536.82182.112191.899.8909442.731466205477.914316523002668637.3614.212731147428801.96577612.15155879316316562.00038116.510.3730139.1134959469110.3915.0868.140.1445.6512.363096076294044720545844388485.9686.247191.07514.97512.2139.06388149.254098.31.993654101.914086.379.840827.817261.053411.427383.7396213.800314.32550.9593362.690252264.462.097142261.842270.925772.413257149.6613.88044.70509548.9250418102588796.46373510404412224526.95536.42182.037192.0210.0319442.733996205477.817316322882301161.5614.156131087431351.95220612.41159872816211201.98550114.840.3730139.0624960686510.3715.0268.390.1445.6812.413096049296344719445843888485.9396.257191.47614.99912.1938.90115142.90OpenBenchmarking.org

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUABC8K16K24K32K40KSE +/- 7.00, N = 3SE +/- 6.25, N = 336805.204099.424098.30MIN: 165671. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUABC48121620SE +/- 0.00178, N = 3SE +/- 0.01869, N = 317.215201.981561.99365MIN: 1.971. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUABC7K14K21K28K35KSE +/- 1.65, N = 3SE +/- 3.15, N = 334757.404102.344101.91MIN: 18514.71. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUABC6K12K18K24K30KSE +/- 3.35, N = 3SE +/- 12.29, N = 329463.304095.964086.37MIN: 14417.6MIN: 4045.08MIN: 4011.271. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUABC1428425670SE +/- 0.01269, N = 3SE +/- 0.00524, N = 364.373709.901069.84082MIN: 14.76MIN: 9.49MIN: 9.411. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPUABC1122334455SE +/- 0.00768, N = 3SE +/- 0.10400, N = 346.670107.698907.81726MIN: 16.15MIN: 7.57MIN: 7.61. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPUABC1224364860SE +/- 0.00153, N = 3SE +/- 0.00112, N = 353.701101.047161.05341MIN: 1.031. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUABC1122334455SE +/- 0.00232, N = 3SE +/- 0.00246, N = 346.398301.426411.42738MIN: 1.41. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUABC306090120150SE +/- 0.00457, N = 3SE +/- 0.00589, N = 3113.724003.740523.73962MIN: 51.181. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUABC1122334455SE +/- 0.02, N = 3SE +/- 0.03, N = 350.7013.7613.80MIN: 19.83MIN: 13.16MIN: 13.21. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUABC1122334455SE +/- 0.01, N = 3SE +/- 0.01, N = 347.6914.3314.33MIN: 17.03MIN: 14.19MIN: 14.161. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUABC714212835SE +/- 0.001376, N = 3SE +/- 0.001531, N = 328.3268000.9578030.959336MIN: 0.921. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPUABC1632486480SE +/- 0.00256, N = 3SE +/- 0.00320, N = 371.826802.689662.69025MIN: 3.121. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUABC7K14K21K28K35KSE +/- 0.80, N = 3SE +/- 3.21, N = 334134.502264.902264.46MIN: 18540.81. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUABC714212835SE +/- 0.00109, N = 3SE +/- 0.00196, N = 327.917202.108942.09714MIN: 2.081. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUABC6K12K18K24K30KSE +/- 1.35, N = 3SE +/- 5.98, N = 328488.202264.982261.84MIN: 124191. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUABC5K10K15K20K25KSE +/- 4.51, N = 3SE +/- 3.05, N = 324467.502271.692270.92MIN: 9924.51. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

ONNX Runtime

ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: fcn-resnet101-11 - Device: CPU - Executor: ParallelABC1326395265SE +/- 0.44, N = 3SE +/- 0.17, N = 34456571. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 10 Realtime - Input: Bosphorus 4KABC1632486480SE +/- 0.14, N = 3SE +/- 0.05, N = 361.3872.3872.411. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

ONNX Runtime

ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: super-resolution-10 - Device: CPU - Executor: ParallelABC7001400210028003500SE +/- 4.73, N = 3SE +/- 11.90, N = 32845324732571. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 10 Realtime - Input: Bosphorus 1080pABC306090120150SE +/- 0.90, N = 3SE +/- 1.27, N = 3133.21150.33149.661. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OSPray

OpenBenchmarking.orgItems Per Second, More Is BetterOSPray 2.9Benchmark: particle_volume/scivis/real_timeABC48121620SE +/- 0.01, N = 3SE +/- 0.02, N = 312.3213.8913.88

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 4KABC1.05752.1153.17254.235.2875SE +/- 0.00, N = 3SE +/- 0.00, N = 34.174.704.701. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OSPray Studio

Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 2 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path TracerABC12K24K36K48K60KSE +/- 44.19, N = 3SE +/- 27.83, N = 35729950897509541. (CXX) g++ options: -O3 -lm -ldl

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 4KABC3691215SE +/- 0.00, N = 3SE +/- 0.07, N = 38.008.998.921. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OSPray Studio

Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 1 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path TracerABC12K24K36K48K60KSE +/- 21.38, N = 3SE +/- 272.17, N = 35635250151504181. (CXX) g++ options: -O3 -lm -ldl

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 1 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path TracerABC20K40K60K80K100KSE +/- 146.65, N = 3SE +/- 84.91, N = 31151451025481025881. (CXX) g++ options: -O3 -lm -ldl

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 1.0Video Input: Summer Nature 1080pABC2004006008001000SE +/- 1.98, N = 3SE +/- 3.38, N = 3710.02795.53796.461. (CC) gcc options: -pthread -lm

OSPray Studio

Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 3 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path TracerABC9001800270036004500SE +/- 3.53, N = 3SE +/- 2.85, N = 34189374037351. (CXX) g++ options: -O3 -lm -ldl

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 2 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path TracerABC20K40K60K80K100KSE +/- 167.02, N = 3SE +/- 63.66, N = 31166891041241040441. (CXX) g++ options: -O3 -lm -ldl

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 3 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path TracerABC30K60K90K120K150KSE +/- 170.37, N = 3SE +/- 103.73, N = 31369291223661222451. (CXX) g++ options: -O3 -lm -ldl

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 1080pABC612182430SE +/- 0.01, N = 3SE +/- 0.00, N = 324.1126.9726.951. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 1.0Video Input: Chimera 1080p 10-bitABC120240360480600SE +/- 0.72, N = 3SE +/- 0.74, N = 3481.75536.82536.421. (CC) gcc options: -pthread -lm

OSPray

OpenBenchmarking.orgItems Per Second, More Is BetterOSPray 2.9Benchmark: particle_volume/pathtracer/real_timeABC4080120160200SE +/- 0.02, N = 3SE +/- 0.10, N = 3163.59182.11182.04

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 1.0Video Input: Summer Nature 4KABC4080120160200SE +/- 0.26, N = 3SE +/- 0.29, N = 3172.71191.89192.021. (CC) gcc options: -pthread -lm

Parallel BZIP2 Compression

This test measures the time needed to compress a file (FreeBSD-13.0-RELEASE-amd64-memstick.img) using Parallel BZIP2 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterParallel BZIP2 Compression 1.1.13FreeBSD-13.0-RELEASE-amd64-memstick.img CompressionABC3691215SE +/- 0.106, N = 3SE +/- 0.023, N = 310.9559.89010.0311. (CXX) g++ options: -O2 -pthread -lbz2 -lpthread

ONNX Runtime

ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: ArcFace ResNet-100 - Device: CPU - Executor: ParallelABC2004006008001000SE +/- 3.33, N = 3SE +/- 2.96, N = 38539449441. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OSPray

OpenBenchmarking.orgItems Per Second, More Is BetterOSPray 2.9Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_timeABC0.61511.23021.84532.46043.0755SE +/- 0.00279, N = 3SE +/- 0.00158, N = 32.472222.731462.73399

OSPray Studio

Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 3 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path TracerABC15K30K45K60K75KSE +/- 147.67, N = 3SE +/- 140.67, N = 36856862054620541. (CXX) g++ options: -O3 -lm -ldl

Timed Wasmer Compilation

This test times how long it takes to compile Wasmer. Wasmer is written in the Rust programming language and is a WebAssembly runtime implementation that supports WASI and EmScripten. This test profile builds Wasmer with the Cranelift and Singlepast compiler features enabled. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Wasmer Compilation 2.2Time To CompileABC20406080100SE +/- 0.11, N = 3SE +/- 0.23, N = 385.4977.9177.821. (CC) gcc options: -m64 -ldl -lxkbcommon -lgcc_s -lutil -lrt -lpthread -lm -lc -pie -nodefaultlibs

OSPray Studio

Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 2 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path TracerABC7001400210028003500SE +/- 0.58, N = 3SE +/- 0.33, N = 33471316531631. (CXX) g++ options: -O3 -lm -ldl

Java JMH

This very basic test profile runs the stock benchmark of the Java JMH benchmark via Maven. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps/s, More Is BetterJava JMHThroughputABC5000M10000M15000M20000M25000M20972409216.5023002668637.3622882301161.56

OSPray

OpenBenchmarking.orgItems Per Second, More Is BetterOSPray 2.9Benchmark: particle_volume/ao/real_timeABC48121620SE +/- 0.03, N = 3SE +/- 0.04, N = 312.9714.2114.16

OSPray Studio

Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 1 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path TracerABC7001400210028003500SE +/- 7.42, N = 3SE +/- 10.02, N = 33398311431081. (CXX) g++ options: -O3 -lm -ldl

Facebook RocksDB

This is a benchmark of Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.0.1Test: Update RandomABC160K320K480K640K800KSE +/- 1507.03, N = 3SE +/- 847.59, N = 36797807428807431351. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OSPray

OpenBenchmarking.orgItems Per Second, More Is BetterOSPray 2.9Benchmark: gravity_spheres_volume/dim_512/scivis/real_timeABC0.44230.88461.32691.76922.2115SE +/- 0.00094, N = 3SE +/- 0.00260, N = 31.799131.965771.95220

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 1.0Video Input: Chimera 1080pABC130260390520650SE +/- 1.17, N = 3SE +/- 1.26, N = 3561.82612.15612.411. (CC) gcc options: -pthread -lm

Facebook RocksDB

This is a benchmark of Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.0.1Test: Read While WritingABC300K600K900K1200K1500KSE +/- 7170.07, N = 3SE +/- 12505.02, N = 31467166155879315987281. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.0.1Test: Read Random Write RandomABC300K600K900K1200K1500KSE +/- 8702.90, N = 3SE +/- 1481.37, N = 31497554163165616211201. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OSPray

OpenBenchmarking.orgItems Per Second, More Is BetterOSPray 2.9Benchmark: gravity_spheres_volume/dim_512/ao/real_timeABC0.45010.90021.35031.80042.2505SE +/- 0.00584, N = 3SE +/- 0.00129, N = 31.838042.000381.98550

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 8 Realtime - Input: Bosphorus 1080pABC306090120150SE +/- 0.41, N = 3SE +/- 0.29, N = 3107.06116.51114.841. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 1080pABC0.08330.16660.24990.33320.4165SE +/- 0.00, N = 3SE +/- 0.00, N = 30.340.370.371. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

ONNX Runtime

ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: yolov4 - Device: CPU - Executor: ParallelABC70140210280350SE +/- 0.88, N = 3SE +/- 1.36, N = 32773013011. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

Timed MPlayer Compilation

This test times how long it takes to build the MPlayer open-source media player program. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed MPlayer Compilation 1.5Time To CompileABC1020304050SE +/- 0.09, N = 3SE +/- 0.08, N = 342.4339.1139.06

Facebook RocksDB

This is a benchmark of Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.0.1Test: Random ReadABC11M22M33M44M55MSE +/- 269279.25, N = 3SE +/- 309988.44, N = 34569720649594691496068651. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 1080pABC3691215SE +/- 0.01, N = 3SE +/- 0.04, N = 39.5810.3910.371. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 6 Realtime - Input: Bosphorus 4KABC48121620SE +/- 0.13, N = 15SE +/- 0.07, N = 313.9115.0815.021. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 9 Realtime - Input: Bosphorus 4KABC1530456075SE +/- 0.12, N = 3SE +/- 0.04, N = 363.1668.1468.391. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 4KABC0.03150.0630.09450.1260.1575SE +/- 0.00, N = 3SE +/- 0.00, N = 30.130.140.141. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 8 Realtime - Input: Bosphorus 4KABC1020304050SE +/- 0.02, N = 3SE +/- 0.07, N = 342.4745.6545.681. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 6 Realtime - Input: Bosphorus 1080pABC3691215SE +/- 0.04, N = 3SE +/- 0.10, N = 811.6712.3612.411. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

ONNX Runtime

ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: yolov4 - Device: CPU - Executor: StandardABC70140210280350SE +/- 0.17, N = 3SE +/- 0.17, N = 32923093091. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: GPT-2 - Device: CPU - Executor: ParallelABC13002600390052006500SE +/- 5.13, N = 3SE +/- 8.66, N = 35745607660491. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: super-resolution-10 - Device: CPU - Executor: StandardABC6001200180024003000SE +/- 12.08, N = 3SE +/- 0.67, N = 32822294029631. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: fcn-resnet101-11 - Device: CPU - Executor: StandardABC1020304050SE +/- 0.00, N = 3SE +/- 0.00, N = 34244441. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: GPT-2 - Device: CPU - Executor: StandardABC15003000450060007500SE +/- 6.71, N = 3SE +/- 2.52, N = 36878720571941. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: bertsquad-12 - Device: CPU - Executor: StandardABC100200300400500SE +/- 0.17, N = 3SE +/- 0.17, N = 34394584581. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: bertsquad-12 - Device: CPU - Executor: ParallelABC100200300400500SE +/- 4.83, N = 12SE +/- 3.68, N = 124274434381. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: ArcFace ResNet-100 - Device: CPU - Executor: StandardABC2004006008001000SE +/- 0.00, N = 3SE +/- 0.17, N = 38698848841. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.10Encoder Speed: 2BCIntel Core i5-1240020406080100SE +/- 0.15, N = 3SE +/- 0.30, N = 385.9785.9485.161. (CXX) g++ options: -O3 -fPIC -lm

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.10Encoder Speed: 10, LosslessBCIntel Core i5-12400246810SE +/- 0.026, N = 3SE +/- 0.029, N = 36.2476.2576.2001. (CXX) g++ options: -O3 -fPIC -lm

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.10Encoder Speed: 0BCIntel Core i5-124004080120160200SE +/- 0.09, N = 3SE +/- 0.20, N = 3191.08191.48189.741. (CXX) g++ options: -O3 -fPIC -lm

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.10Encoder Speed: 6, LosslessBCIntel Core i5-1240048121620SE +/- 0.03, N = 3SE +/- 0.03, N = 314.9815.0014.921. (CXX) g++ options: -O3 -fPIC -lm

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.10Encoder Speed: 6BCIntel Core i5-124003691215SE +/- 0.03, N = 3SE +/- 0.02, N = 312.2112.1912.211. (CXX) g++ options: -O3 -fPIC -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPUABC4080120160200SE +/- 0.17309, N = 15SE +/- 0.13280, N = 15162.853009.063888.90115MIN: 6.381. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 9 Realtime - Input: Bosphorus 1080pABC306090120150SE +/- 2.69, N = 15SE +/- 0.31, N = 3124.60149.25142.901. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

76 Results Shown

oneDNN:
  Recurrent Neural Network Training - bf16bf16bf16 - CPU
  Deconvolution Batch shapes_3d - u8s8f32 - CPU
  Recurrent Neural Network Training - u8s8f32 - CPU
  Recurrent Neural Network Training - f32 - CPU
  IP Shapes 3D - f32 - CPU
  Deconvolution Batch shapes_3d - f32 - CPU
  Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU
  Deconvolution Batch shapes_1d - u8s8f32 - CPU
  IP Shapes 1D - f32 - CPU
  Convolution Batch Shapes Auto - u8s8f32 - CPU
  Convolution Batch Shapes Auto - f32 - CPU
  IP Shapes 1D - u8s8f32 - CPU
  Matrix Multiply Batch Shapes Transformer - f32 - CPU
  Recurrent Neural Network Inference - f32 - CPU
  IP Shapes 3D - u8s8f32 - CPU
  Recurrent Neural Network Inference - u8s8f32 - CPU
  Recurrent Neural Network Inference - bf16bf16bf16 - CPU
ONNX Runtime
AOM AV1
ONNX Runtime
AOM AV1
OSPray
AOM AV1
OSPray Studio
AOM AV1
OSPray Studio:
  1 - 1080p - 16 - Path Tracer
  1 - 1080p - 32 - Path Tracer
dav1d
OSPray Studio:
  3 - 1080p - 1 - Path Tracer
  2 - 1080p - 32 - Path Tracer
  3 - 1080p - 32 - Path Tracer
AOM AV1
dav1d
OSPray
dav1d
Parallel BZIP2 Compression
ONNX Runtime
OSPray
OSPray Studio
Timed Wasmer Compilation
OSPray Studio
Java JMH
OSPray
OSPray Studio
Facebook RocksDB
OSPray
dav1d
Facebook RocksDB:
  Read While Writing
  Read Rand Write Rand
OSPray
AOM AV1:
  Speed 8 Realtime - Bosphorus 1080p
  Speed 0 Two-Pass - Bosphorus 1080p
ONNX Runtime
Timed MPlayer Compilation
Facebook RocksDB
AOM AV1:
  Speed 4 Two-Pass - Bosphorus 1080p
  Speed 6 Realtime - Bosphorus 4K
  Speed 9 Realtime - Bosphorus 4K
  Speed 0 Two-Pass - Bosphorus 4K
  Speed 8 Realtime - Bosphorus 4K
  Speed 6 Realtime - Bosphorus 1080p
ONNX Runtime:
  yolov4 - CPU - Standard
  GPT-2 - CPU - Parallel
  super-resolution-10 - CPU - Standard
  fcn-resnet101-11 - CPU - Standard
  GPT-2 - CPU - Standard
  bertsquad-12 - CPU - Standard
  bertsquad-12 - CPU - Parallel
  ArcFace ResNet-100 - CPU - Standard
libavif avifenc:
  2
  10, Lossless
  0
  6, Lossless
  6
oneDNN
AOM AV1