core-i5-12400-april

Intel Core i5-12400 testing with a ASRock B660M-HDV (3.02 BIOS) and llvmpipe on Ubuntu 22.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2204077-NE-COREI512423
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results
Show Result Confidence Charts
Allow Limiting Results To Certain Suite(s)

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Disable Color Branding
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Toggle/Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
A
April 06 2022
  1 Hour, 59 Minutes
Intel Core i5-12400
April 06 2022
  6 Minutes
B
April 06 2022
  5 Hours, 40 Minutes
C
April 07 2022
  5 Hours, 38 Minutes
Invert Behavior (Only Show Selected Data)
  3 Hours, 21 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


core-i5-12400-april ProcessorMotherboardChipsetMemoryDiskGraphicsAudioNetworkOSKernelDesktopDisplay ServerOpenGLVulkanCompilerFile-SystemScreen ResolutionAIntel Core i5-12400BCIntel Core i5-12400 @ 5.60GHz (6 Cores / 12 Threads)ASRock B660M-HDV (3.02 BIOS)Intel Device 7aa716GB512GB SabrentllvmpipeRealtek ALC897IntelUbuntu 22.045.15.0-18-generic (x86_64)GNOME Shell 41.3X Server 1.20.144.5 Mesa 21.2.2 (LLVM 12.0.1 256 bits)1.1.182GCC 11.2.0ext41920x10804.5 Mesa 22.0.1 (LLVM 13.0.1 256 bits)1.2.204OpenBenchmarking.orgKernel Details- Transparent Huge Pages: madviseCompiler Details- A: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-iOLsLC/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-iOLsLC/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Intel Core i5-12400: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - B: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - C: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- A: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.7- Intel Core i5-12400: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.9- B: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.9- C: Scaling Governor: intel_pstate powersave (EPP: balance_performance) - CPU Microcode: 0x18 - Thermald 2.4.9Java Details- A, B, C: OpenJDK Runtime Environment (build 11.0.14.1+1-Ubuntu-0ubuntu1)Python Details- A: Python 3.9.12- B: Python 3.10.4- C: Python 3.10.4Security Details- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected

core-i5-12400-april onnx: bertsquad-12 - CPU - Parallelospray: particle_volume/scivis/real_timeospray: particle_volume/pathtracer/real_timeospray: particle_volume/ao/real_timejava-jmh: Throughputavifenc: 0ospray-studio: 2 - 1080p - 16 - Path Tracerospray-studio: 1 - 1080p - 16 - Path Traceraom-av1: Speed 4 Two-Pass - Bosphorus 4Kospray: gravity_spheres_volume/dim_512/scivis/real_timeospray: gravity_spheres_volume/dim_512/ao/real_timeaom-av1: Speed 0 Two-Pass - Bosphorus 4Kospray-studio: 3 - 1080p - 1 - Path Tracerospray-studio: 1 - 1080p - 1 - Path Tracerospray-studio: 2 - 1080p - 1 - Path Tracerospray: gravity_spheres_volume/dim_512/pathtracer/real_timeospray-studio: 3 - 1080p - 32 - Path Traceronnx: fcn-resnet101-11 - CPU - Parallelonnx: fcn-resnet101-11 - CPU - Standardonnx: GPT-2 - CPU - Parallelonnx: yolov4 - CPU - Parallelonnx: ArcFace ResNet-100 - CPU - Parallelonnx: GPT-2 - CPU - Standardonnx: bertsquad-12 - CPU - Standardonnx: ArcFace ResNet-100 - CPU - Standardonnx: yolov4 - CPU - Standardonnx: super-resolution-10 - CPU - Standardonnx: super-resolution-10 - CPU - Parallelospray-studio: 2 - 1080p - 32 - Path Tracerospray-studio: 1 - 1080p - 32 - Path Traceraom-av1: Speed 6 Realtime - Bosphorus 4Konednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUonednn: Deconvolution Batch shapes_1d - f32 - CPUavifenc: 2aom-av1: Speed 6 Realtime - Bosphorus 1080paom-av1: Speed 6 Two-Pass - Bosphorus 4Kbuild-wasmer: Time To Compileospray-studio: 3 - 1080p - 16 - Path Traceraom-av1: Speed 4 Two-Pass - Bosphorus 1080procksdb: Update Randrocksdb: Read Rand Write Randrocksdb: Read While Writingrocksdb: Rand Readaom-av1: Speed 0 Two-Pass - Bosphorus 1080pbuild-mplayer: Time To Compileaom-av1: Speed 6 Two-Pass - Bosphorus 1080ponednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUdav1d: Summer Nature 4Kdav1d: Chimera 1080p 10-bitonednn: IP Shapes 1D - f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUdav1d: Chimera 1080pavifenc: 6, Losslessaom-av1: Speed 8 Realtime - Bosphorus 4Kavifenc: 6onednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUaom-av1: Speed 9 Realtime - Bosphorus 1080pcompress-pbzip2: FreeBSD-13.0-RELEASE-amd64-memstick.img Compressionaom-av1: Speed 9 Realtime - Bosphorus 4Konednn: IP Shapes 3D - f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUaom-av1: Speed 10 Realtime - Bosphorus 4Kavifenc: 10, Losslessonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUaom-av1: Speed 8 Realtime - Bosphorus 1080pdav1d: Summer Nature 1080paom-av1: Speed 10 Realtime - Bosphorus 1080ponednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: IP Shapes 1D - bf16bf16bf16 - CPUAIntel Core i5-12400BC42712.3203163.59112.968820972409216.50257299563524.171.799131.838040.134189339834712.472221369294442574527785368784398692922822284511668911514513.9136805.234757.429463.334134.528488.224467.5162.85311.67885.485685689.5867978014975541467166456972060.3442.43224.1146.3983172.71481.75113.72428.3268561.8242.4771.826853.7011124.610.95563.1664.373727.917261.3850.702447.6937107.06710.02133.2146.670117.2152189.7485.1614.91812.2076.244313.8893182.11214.212723002668637.36191.07550897501514.701.965772.000380.143740311431652.731461223665644607630194472054588843092940324710412410254815.084099.424102.344095.962264.902264.982271.699.0638885.96812.368.9977.9146205410.3974288016316561558793495946910.3739.11326.971.42641191.89536.823.740520.957803612.1514.97545.6512.2132.689661.04716149.259.89068.149.901062.1089472.386.24713.756514.3318116.51795.53150.337.698901.9815643813.8804182.03714.156122882301161.56191.47650954504184.701.952201.985500.143735310831632.733991222455744604930194471944588843092963325710404410258815.024098.34101.914086.372264.462261.842270.928.9011585.93912.418.9277.8176205410.3774313516211201598728496068650.3739.06226.951.42738192.02536.423.739620.959336612.4114.99945.6812.1932.690251.05341142.9010.03168.399.840822.0971472.416.25713.800314.3255114.84796.46149.667.817261.99365OpenBenchmarking.org

ONNX Runtime

ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: bertsquad-12 - Device: CPU - Executor: ParallelBCA100200300400500SE +/- 4.83, N = 12SE +/- 3.68, N = 124434384271. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OSPray

OpenBenchmarking.orgItems Per Second, More Is BetterOSPray 2.9Benchmark: particle_volume/scivis/real_timeBCA48121620SE +/- 0.01, N = 3SE +/- 0.02, N = 313.8913.8812.32

OpenBenchmarking.orgItems Per Second, More Is BetterOSPray 2.9Benchmark: particle_volume/pathtracer/real_timeBCA4080120160200SE +/- 0.02, N = 3SE +/- 0.10, N = 3182.11182.04163.59

OpenBenchmarking.orgItems Per Second, More Is BetterOSPray 2.9Benchmark: particle_volume/ao/real_timeBCA48121620SE +/- 0.03, N = 3SE +/- 0.04, N = 314.2114.1612.97

Java JMH

This very basic test profile runs the stock benchmark of the Java JMH benchmark via Maven. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps/s, More Is BetterJava JMHThroughputBCA5000M10000M15000M20000M25000M23002668637.3622882301161.5620972409216.50

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.10Encoder Speed: 0Intel Core i5-12400BC4080120160200SE +/- 0.09, N = 3SE +/- 0.20, N = 3189.74191.08191.481. (CXX) g++ options: -O3 -fPIC -lm

OSPray Studio

Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 2 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path TracerBCA12K24K36K48K60KSE +/- 44.19, N = 3SE +/- 27.83, N = 35089750954572991. (CXX) g++ options: -O3 -lm -ldl

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 1 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path TracerBCA12K24K36K48K60KSE +/- 21.38, N = 3SE +/- 272.17, N = 35015150418563521. (CXX) g++ options: -O3 -lm -ldl

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 4KCBA1.05752.1153.17254.235.2875SE +/- 0.00, N = 3SE +/- 0.00, N = 34.704.704.171. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OSPray

OpenBenchmarking.orgItems Per Second, More Is BetterOSPray 2.9Benchmark: gravity_spheres_volume/dim_512/scivis/real_timeBCA0.44230.88461.32691.76922.2115SE +/- 0.00094, N = 3SE +/- 0.00260, N = 31.965771.952201.79913

OpenBenchmarking.orgItems Per Second, More Is BetterOSPray 2.9Benchmark: gravity_spheres_volume/dim_512/ao/real_timeBCA0.45010.90021.35031.80042.2505SE +/- 0.00584, N = 3SE +/- 0.00129, N = 32.000381.985501.83804

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 4KCBA0.03150.0630.09450.1260.1575SE +/- 0.00, N = 3SE +/- 0.00, N = 30.140.140.131. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OSPray Studio

Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 3 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path TracerCBA9001800270036004500SE +/- 2.85, N = 3SE +/- 3.53, N = 33735374041891. (CXX) g++ options: -O3 -lm -ldl

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 1 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path TracerCBA7001400210028003500SE +/- 10.02, N = 3SE +/- 7.42, N = 33108311433981. (CXX) g++ options: -O3 -lm -ldl

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 2 - Resolution: 1080p - Samples Per Pixel: 1 - Renderer: Path TracerCBA7001400210028003500SE +/- 0.33, N = 3SE +/- 0.58, N = 33163316534711. (CXX) g++ options: -O3 -lm -ldl

OSPray

OpenBenchmarking.orgItems Per Second, More Is BetterOSPray 2.9Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_timeCBA0.61511.23021.84532.46043.0755SE +/- 0.00158, N = 3SE +/- 0.00279, N = 32.733992.731462.47222

OSPray Studio

Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 3 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path TracerCBA30K60K90K120K150KSE +/- 103.73, N = 3SE +/- 170.37, N = 31222451223661369291. (CXX) g++ options: -O3 -lm -ldl

ONNX Runtime

ONNX Runtime is developed by Microsoft and partners as a open-source, cross-platform, high performance machine learning inferencing and training accelerator. This test profile runs the ONNX Runtime with various models available from the ONNX Zoo. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: fcn-resnet101-11 - Device: CPU - Executor: ParallelCBA1326395265SE +/- 0.17, N = 3SE +/- 0.44, N = 35756441. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: fcn-resnet101-11 - Device: CPU - Executor: StandardCBA1020304050SE +/- 0.00, N = 3SE +/- 0.00, N = 34444421. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: GPT-2 - Device: CPU - Executor: ParallelBCA13002600390052006500SE +/- 5.13, N = 3SE +/- 8.66, N = 36076604957451. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: yolov4 - Device: CPU - Executor: ParallelCBA70140210280350SE +/- 1.36, N = 3SE +/- 0.88, N = 33013012771. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: ArcFace ResNet-100 - Device: CPU - Executor: ParallelCBA2004006008001000SE +/- 2.96, N = 3SE +/- 3.33, N = 39449448531. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: GPT-2 - Device: CPU - Executor: StandardBCA15003000450060007500SE +/- 6.71, N = 3SE +/- 2.52, N = 37205719468781. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: bertsquad-12 - Device: CPU - Executor: StandardCBA100200300400500SE +/- 0.17, N = 3SE +/- 0.17, N = 34584584391. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: ArcFace ResNet-100 - Device: CPU - Executor: StandardCBA2004006008001000SE +/- 0.17, N = 3SE +/- 0.00, N = 38848848691. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: yolov4 - Device: CPU - Executor: StandardCBA70140210280350SE +/- 0.17, N = 3SE +/- 0.17, N = 33093092921. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: super-resolution-10 - Device: CPU - Executor: StandardCBA6001200180024003000SE +/- 0.67, N = 3SE +/- 12.08, N = 32963294028221. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Minute, More Is BetterONNX Runtime 1.11Model: super-resolution-10 - Device: CPU - Executor: ParallelCBA7001400210028003500SE +/- 11.90, N = 3SE +/- 4.73, N = 33257324728451. (CXX) g++ options: -ffunction-sections -fdata-sections -march=native -mtune=native -O3 -flto -fno-fat-lto-objects -ldl -lrt

OSPray Studio

Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 2 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path TracerCBA20K40K60K80K100KSE +/- 63.66, N = 3SE +/- 167.02, N = 31040441041241166891. (CXX) g++ options: -O3 -lm -ldl

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 1 - Resolution: 1080p - Samples Per Pixel: 32 - Renderer: Path TracerBCA20K40K60K80K100KSE +/- 146.65, N = 3SE +/- 84.91, N = 31025481025881151451. (CXX) g++ options: -O3 -lm -ldl

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 6 Realtime - Input: Bosphorus 4KBCA48121620SE +/- 0.13, N = 15SE +/- 0.07, N = 315.0815.0213.911. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUCBA8K16K24K32K40KSE +/- 6.25, N = 3SE +/- 7.00, N = 34098.304099.4236805.20MIN: 165671. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUCBA7K14K21K28K35KSE +/- 3.15, N = 3SE +/- 1.65, N = 34101.914102.3434757.40MIN: 18514.71. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUCBA6K12K18K24K30KSE +/- 12.29, N = 3SE +/- 3.35, N = 34086.374095.9629463.30MIN: 4011.27MIN: 4045.08MIN: 14417.61. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUCBA7K14K21K28K35KSE +/- 3.21, N = 3SE +/- 0.80, N = 32264.462264.9034134.50MIN: 18540.81. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUCBA6K12K18K24K30KSE +/- 5.98, N = 3SE +/- 1.35, N = 32261.842264.9828488.20MIN: 124191. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUCBA5K10K15K20K25KSE +/- 3.05, N = 3SE +/- 4.51, N = 32270.922271.6924467.50MIN: 9924.51. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPUCBA4080120160200SE +/- 0.13280, N = 15SE +/- 0.17309, N = 158.901159.06388162.85300MIN: 6.381. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.10Encoder Speed: 2Intel Core i5-12400CB20406080100SE +/- 0.30, N = 3SE +/- 0.15, N = 385.1685.9485.971. (CXX) g++ options: -O3 -fPIC -lm

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 6 Realtime - Input: Bosphorus 1080pCBA3691215SE +/- 0.10, N = 8SE +/- 0.04, N = 312.4112.3611.671. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 4KBCA3691215SE +/- 0.00, N = 3SE +/- 0.07, N = 38.998.928.001. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

Timed Wasmer Compilation

This test times how long it takes to compile Wasmer. Wasmer is written in the Rust programming language and is a WebAssembly runtime implementation that supports WASI and EmScripten. This test profile builds Wasmer with the Cranelift and Singlepast compiler features enabled. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Wasmer Compilation 2.2Time To CompileCBA20406080100SE +/- 0.23, N = 3SE +/- 0.11, N = 377.8277.9185.491. (CC) gcc options: -m64 -ldl -lxkbcommon -lgcc_s -lutil -lrt -lpthread -lm -lc -pie -nodefaultlibs

OSPray Studio

Intel OSPray Studio is an open-source, interactive visualization and ray-tracing software package. OSPray Studio makes use of Intel OSPray, a portable ray-tracing engine for high-performance, high-fidelity visualizations. OSPray builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOSPray Studio 0.10Camera: 3 - Resolution: 1080p - Samples Per Pixel: 16 - Renderer: Path TracerBCA15K30K45K60K75KSE +/- 147.67, N = 3SE +/- 140.67, N = 36205462054685681. (CXX) g++ options: -O3 -lm -ldl

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 1080pBCA3691215SE +/- 0.01, N = 3SE +/- 0.04, N = 310.3910.379.581. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

Facebook RocksDB

This is a benchmark of Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.0.1Test: Update RandomCBA160K320K480K640K800KSE +/- 847.59, N = 3SE +/- 1507.03, N = 37431357428806797801. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.0.1Test: Read Random Write RandomBCA300K600K900K1200K1500KSE +/- 8702.90, N = 3SE +/- 1481.37, N = 31631656162112014975541. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.0.1Test: Read While WritingCBA300K600K900K1200K1500KSE +/- 12505.02, N = 3SE +/- 7170.07, N = 31598728155879314671661. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.0.1Test: Random ReadCBA11M22M33M44M55MSE +/- 309988.44, N = 3SE +/- 269279.25, N = 34960686549594691456972061. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 1080pCBA0.08330.16660.24990.33320.4165SE +/- 0.00, N = 3SE +/- 0.00, N = 30.370.370.341. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

Timed MPlayer Compilation

This test times how long it takes to build the MPlayer open-source media player program. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed MPlayer Compilation 1.5Time To CompileCBA1020304050SE +/- 0.08, N = 3SE +/- 0.09, N = 339.0639.1142.43

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 1080pBCA612182430SE +/- 0.01, N = 3SE +/- 0.00, N = 326.9726.9524.111. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUBCA1122334455SE +/- 0.00232, N = 3SE +/- 0.00246, N = 31.426411.4273846.39830MIN: 1.41. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 1.0Video Input: Summer Nature 4KCBA4080120160200SE +/- 0.29, N = 3SE +/- 0.26, N = 3192.02191.89172.711. (CC) gcc options: -pthread -lm

OpenBenchmarking.orgFPS, More Is Betterdav1d 1.0Video Input: Chimera 1080p 10-bitBCA120240360480600SE +/- 0.72, N = 3SE +/- 0.74, N = 3536.82536.42481.751. (CC) gcc options: -pthread -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUCBA306090120150SE +/- 0.00589, N = 3SE +/- 0.00457, N = 33.739623.74052113.72400MIN: 51.181. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUBCA714212835SE +/- 0.001376, N = 3SE +/- 0.001531, N = 30.9578030.95933628.326800MIN: 0.921. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 1.0Video Input: Chimera 1080pCBA130260390520650SE +/- 1.26, N = 3SE +/- 1.17, N = 3612.41612.15561.821. (CC) gcc options: -pthread -lm

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.10Encoder Speed: 6, LosslessIntel Core i5-12400BC48121620SE +/- 0.03, N = 3SE +/- 0.03, N = 314.9214.9815.001. (CXX) g++ options: -O3 -fPIC -lm

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 8 Realtime - Input: Bosphorus 4KCBA1020304050SE +/- 0.07, N = 3SE +/- 0.02, N = 345.6845.6542.471. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.10Encoder Speed: 6CIntel Core i5-12400B3691215SE +/- 0.02, N = 3SE +/- 0.03, N = 312.1912.2112.211. (CXX) g++ options: -O3 -fPIC -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPUBCA1632486480SE +/- 0.00256, N = 3SE +/- 0.00320, N = 32.689662.6902571.82680MIN: 3.121. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPUBCA1224364860SE +/- 0.00153, N = 3SE +/- 0.00112, N = 31.047161.0534153.70110MIN: 1.031. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 9 Realtime - Input: Bosphorus 1080pBCA306090120150SE +/- 2.69, N = 15SE +/- 0.31, N = 3149.25142.90124.601. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

Parallel BZIP2 Compression

This test measures the time needed to compress a file (FreeBSD-13.0-RELEASE-amd64-memstick.img) using Parallel BZIP2 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterParallel BZIP2 Compression 1.1.13FreeBSD-13.0-RELEASE-amd64-memstick.img CompressionBCA3691215SE +/- 0.106, N = 3SE +/- 0.023, N = 39.89010.03110.9551. (CXX) g++ options: -O2 -pthread -lbz2 -lpthread

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 9 Realtime - Input: Bosphorus 4KCBA1530456075SE +/- 0.04, N = 3SE +/- 0.12, N = 368.3968.1463.161. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUCBA1428425670SE +/- 0.00524, N = 3SE +/- 0.01269, N = 39.840829.9010664.37370MIN: 9.41MIN: 9.49MIN: 14.761. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUCBA714212835SE +/- 0.00196, N = 3SE +/- 0.00109, N = 32.097142.1089427.91720MIN: 2.081. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 10 Realtime - Input: Bosphorus 4KCBA1632486480SE +/- 0.05, N = 3SE +/- 0.14, N = 372.4172.3861.381. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

libavif avifenc

This is a test of the AOMedia libavif library testing the encoding of a JPEG image to AV1 Image Format (AVIF). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 0.10Encoder Speed: 10, LosslessIntel Core i5-12400BC246810SE +/- 0.026, N = 3SE +/- 0.029, N = 36.2006.2476.2571. (CXX) g++ options: -O3 -fPIC -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUBCA1122334455SE +/- 0.02, N = 3SE +/- 0.03, N = 313.7613.8050.70MIN: 13.16MIN: 13.2MIN: 19.831. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUCBA1122334455SE +/- 0.01, N = 3SE +/- 0.01, N = 314.3314.3347.69MIN: 14.16MIN: 14.19MIN: 17.031. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 8 Realtime - Input: Bosphorus 1080pBCA306090120150SE +/- 0.41, N = 3SE +/- 0.29, N = 3116.51114.84107.061. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

dav1d

Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is Betterdav1d 1.0Video Input: Summer Nature 1080pCBA2004006008001000SE +/- 3.38, N = 3SE +/- 1.98, N = 3796.46795.53710.021. (CC) gcc options: -pthread -lm

AOM AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.3Encoder Mode: Speed 10 Realtime - Input: Bosphorus 1080pBCA306090120150SE +/- 0.90, N = 3SE +/- 1.27, N = 3150.33149.66133.211. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPUBCA1122334455SE +/- 0.00768, N = 3SE +/- 0.10400, N = 37.698907.8172646.67010MIN: 7.57MIN: 7.6MIN: 16.151. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUBCA48121620SE +/- 0.00178, N = 3SE +/- 0.01869, N = 31.981561.9936517.21520MIN: 1.971. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -ldl

Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

76 Results Shown

ONNX Runtime
OSPray:
  particle_volume/scivis/real_time
  particle_volume/pathtracer/real_time
  particle_volume/ao/real_time
Java JMH
libavif avifenc
OSPray Studio:
  2 - 1080p - 16 - Path Tracer
  1 - 1080p - 16 - Path Tracer
AOM AV1
OSPray:
  gravity_spheres_volume/dim_512/scivis/real_time
  gravity_spheres_volume/dim_512/ao/real_time
AOM AV1
OSPray Studio:
  3 - 1080p - 1 - Path Tracer
  1 - 1080p - 1 - Path Tracer
  2 - 1080p - 1 - Path Tracer
OSPray
OSPray Studio
ONNX Runtime:
  fcn-resnet101-11 - CPU - Parallel
  fcn-resnet101-11 - CPU - Standard
  GPT-2 - CPU - Parallel
  yolov4 - CPU - Parallel
  ArcFace ResNet-100 - CPU - Parallel
  GPT-2 - CPU - Standard
  bertsquad-12 - CPU - Standard
  ArcFace ResNet-100 - CPU - Standard
  yolov4 - CPU - Standard
  super-resolution-10 - CPU - Standard
  super-resolution-10 - CPU - Parallel
OSPray Studio:
  2 - 1080p - 32 - Path Tracer
  1 - 1080p - 32 - Path Tracer
AOM AV1
oneDNN:
  Recurrent Neural Network Training - bf16bf16bf16 - CPU
  Recurrent Neural Network Training - u8s8f32 - CPU
  Recurrent Neural Network Training - f32 - CPU
  Recurrent Neural Network Inference - f32 - CPU
  Recurrent Neural Network Inference - u8s8f32 - CPU
  Recurrent Neural Network Inference - bf16bf16bf16 - CPU
  Deconvolution Batch shapes_1d - f32 - CPU
libavif avifenc
AOM AV1:
  Speed 6 Realtime - Bosphorus 1080p
  Speed 6 Two-Pass - Bosphorus 4K
Timed Wasmer Compilation
OSPray Studio
AOM AV1
Facebook RocksDB:
  Update Rand
  Read Rand Write Rand
  Read While Writing
  Rand Read
AOM AV1
Timed MPlayer Compilation
AOM AV1
oneDNN
dav1d:
  Summer Nature 4K
  Chimera 1080p 10-bit
oneDNN:
  IP Shapes 1D - f32 - CPU
  IP Shapes 1D - u8s8f32 - CPU
dav1d
libavif avifenc
AOM AV1
libavif avifenc
oneDNN:
  Matrix Multiply Batch Shapes Transformer - f32 - CPU
  Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU
AOM AV1
Parallel BZIP2 Compression
AOM AV1
oneDNN:
  IP Shapes 3D - f32 - CPU
  IP Shapes 3D - u8s8f32 - CPU
AOM AV1
libavif avifenc
oneDNN:
  Convolution Batch Shapes Auto - u8s8f32 - CPU
  Convolution Batch Shapes Auto - f32 - CPU
AOM AV1
dav1d
AOM AV1
oneDNN:
  Deconvolution Batch shapes_3d - f32 - CPU
  Deconvolution Batch shapes_3d - u8s8f32 - CPU