extra tests2

Tests for a future article. AMD EPYC 9124 16-Core testing with a Supermicro H13SSW (1.1 BIOS) and astdrmfb on AlmaLinux 9.2 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2310228-NE-EXTRATEST37
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results

Limit displaying results to tests within:

CPU Massive 5 Tests
Creator Workloads 9 Tests
Database Test Suite 2 Tests
Game Development 3 Tests
HPC - High Performance Computing 6 Tests
Java Tests 2 Tests
Machine Learning 3 Tests
Multi-Core 10 Tests
Intel oneAPI 6 Tests
OpenMPI Tests 5 Tests
Python Tests 2 Tests
Renderers 2 Tests
Server 2 Tests
Server CPU Tests 4 Tests
Common Workstation Benchmarks 2 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Additional Graphs

Show Perf Per Core/Thread Calculation Graphs Where Applicable
Show Perf Per Clock Calculation Graphs Where Applicable

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs
Condense Test Profiles With Multiple Version Results Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
a
October 07 2023
  5 Hours, 38 Minutes
b
October 07 2023
  5 Hours, 6 Minutes
c
October 07 2023
  4 Hours, 48 Minutes
d
October 20 2023
  5 Hours, 58 Minutes
e
October 20 2023
  6 Hours, 40 Minutes
f
October 21 2023
  6 Hours, 19 Minutes
g
October 22 2023
  5 Hours, 59 Minutes
Invert Hiding All Results Option
  5 Hours, 47 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


extra tests2ProcessorMotherboardMemoryDiskGraphicsOSKernelCompilerFile-SystemScreen Resolutionabcdefg2 x AMD EPYC 9254 24-Core @ 2.90GHz (48 Cores / 96 Threads)Supermicro H13DSH (1.5 BIOS)24 x 32 GB DDR5-4800MT/s Samsung M321R4GA3BB6-CQKET2 x 1920GB SAMSUNG MZQL21T9HCJR-00A07astdrmfbAlmaLinux 9.25.14.0-284.25.1.el9_2.x86_64 (x86_64)GCC 11.3.1 20221121ext41024x768AMD EPYC 9124 16-Core @ 3.00GHz (16 Cores / 32 Threads)Supermicro H13SSW (1.1 BIOS)12 x 64 GB DDR5-4800MT/s HMCG94MEBRA123NOpenBenchmarking.orgKernel Details- Transparent Huge Pages: alwaysCompiler Details- --build=x86_64-redhat-linux --disable-libunwind-exceptions --enable-__cxa_atexit --enable-bootstrap --enable-cet --enable-checking=release --enable-gnu-indirect-function --enable-gnu-unique-object --enable-host-bind-now --enable-host-pie --enable-initfini-array --enable-languages=c,c++,fortran,lto --enable-link-serialization=1 --enable-multilib --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --mandir=/usr/share/man --with-arch_32=x86-64 --with-arch_64=x86-64-v2 --with-build-config=bootstrap-lto --with-gcc-major-version-only --with-linker-hash-style=gnu --with-tune=generic --without-cuda-driver --without-isl Processor Details- a: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa10113e- b: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa10113e- c: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa10113e- d: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa101111- e: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa101111- f: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa101111- g: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa101111Java Details- OpenJDK Runtime Environment (Red_Hat-11.0.20.0.8-1) (build 11.0.20+8-LTS)Python Details- Python 3.9.16Security Details- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected

abcdefgResult OverviewPhoronix Test Suite100%143%187%230%273%BlenderBRL-CADIntel Open Image DenoiseEmbreeSPECFEM3DOSPRayTimed Linux Kernel CompilationRemhosOpenVINONeural Magic DeepSparseTiDB Community ServerApache HadoopApache CassandraLiquid-DSPSVT-AV1nekRS

extra tests2brl-cad: VGR Performance Metricopenvkl: vklBenchmarkCPU Scalartidb: oltp_read_write - 16tidb: oltp_read_write - 32tidb: oltp_update_non_index - 1tidb: oltp_read_write - 64openvkl: vklBenchmarkCPU ISPCtidb: oltp_update_non_index - 64tidb: oltp_read_write - 1tidb: oltp_update_index - 16tidb: oltp_point_select - 1tidb: oltp_point_select - 128tidb: oltp_point_select - 32tidb: oltp_point_select - 64tidb: oltp_update_index - 32tidb: oltp_update_index - 128tidb: oltp_update_index - 1tidb: oltp_read_write - 128blender: Barbershop - CPU-Onlytidb: oltp_update_non_index - 16tidb: oltp_update_non_index - 32tidb: oltp_update_index - 64tidb: oltp_update_non_index - 128tidb: oltp_point_select - 16blender: Pabellon Barcelona - CPU-Onlyospray: particle_volume/scivis/real_timeospray: particle_volume/pathtracer/real_timenekrs: TurboPipe Periodicblender: Classroom - CPU-Onlycassandra: Writesospray: particle_volume/ao/real_timedeepsparse: BERT-Large, NLP Question Answering - Asynchronous Multi-Streamdeepsparse: BERT-Large, NLP Question Answering - Asynchronous Multi-Streameasywave: e2Asean Grid + BengkuluSept2007 Source - 2400nekrs: Kershawoidn: RTLightmap.hdr.4096x4096 - CPU-Onlyonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUblender: Fishy Cat - CPU-Onlyopenvino: Face Detection FP16 - CPUopenvino: Face Detection FP16 - CPUopenvino: Face Detection FP16-INT8 - CPUopenvino: Face Detection FP16-INT8 - CPUospray: gravity_spheres_volume/dim_512/scivis/real_timeospray: gravity_spheres_volume/dim_512/ao/real_timeopenvino: Person Detection FP16 - CPUopenvino: Person Detection FP16 - CPUopenvino: Person Detection FP32 - CPUopenvino: Person Detection FP32 - CPUopenvino: Machine Translation EN To DE FP16 - CPUopenvino: Machine Translation EN To DE FP16 - CPUopenvino: Road Segmentation ADAS FP16-INT8 - CPUopenvino: Road Segmentation ADAS FP16-INT8 - CPUopenvino: Person Vehicle Bike Detection FP16 - CPUopenvino: Person Vehicle Bike Detection FP16 - CPUopenvino: Road Segmentation ADAS FP16 - CPUopenvino: Road Segmentation ADAS FP16 - CPUopenvino: Face Detection Retail FP16-INT8 - CPUopenvino: Face Detection Retail FP16-INT8 - CPUopenvino: Handwritten English Recognition FP16-INT8 - CPUopenvino: Handwritten English Recognition FP16-INT8 - CPUopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUopenvino: Handwritten English Recognition FP16 - CPUopenvino: Handwritten English Recognition FP16 - CPUopenvino: Vehicle Detection FP16-INT8 - CPUopenvino: Vehicle Detection FP16-INT8 - CPUopenvino: Age Gender Recognition Retail 0013 FP16 - CPUopenvino: Age Gender Recognition Retail 0013 FP16 - CPUopenvino: Weld Porosity Detection FP16 - CPUopenvino: Weld Porosity Detection FP16 - CPUopenvino: Vehicle Detection FP16 - CPUopenvino: Vehicle Detection FP16 - CPUopenvino: Weld Porosity Detection FP16-INT8 - CPUopenvino: Weld Porosity Detection FP16-INT8 - CPUopenvino: Face Detection Retail FP16 - CPUopenvino: Face Detection Retail FP16 - CPUospray: gravity_spheres_volume/dim_512/pathtracer/real_timespecfem3d: Layered Halfspacespecfem3d: Water-layered Halfspacedeepsparse: BERT-Large, NLP Question Answering, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: BERT-Large, NLP Question Answering, Sparse INT8 - Asynchronous Multi-Streamblender: BMW27 - CPU-Onlyhadoop: Rename - 100 - 1000000deepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Streamdeepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Streamhadoop: Rename - 50 - 1000000deepsparse: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Asynchronous Multi-Streamhadoop: Delete - 100 - 1000000deepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Streamdeepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Streamhadoop: Delete - 50 - 1000000deepsparse: NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Streamdeepsparse: NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Streamdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Streamdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Streamdeepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Asynchronous Multi-Streamdeepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Asynchronous Multi-Streamoidn: RTLightmap.hdr.4096x4096 - CPU-Onlybuild-linux-kernel: defconfigdeepsparse: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Asynchronous Multi-Streamdeepsparse: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Asynchronous Multi-Streamhadoop: Open - 100 - 1000000embree: Pathtracer - Asian Dragon Objoidn: RT.hdr_alb_nrm.3840x2160 - CPU-Onlyoidn: RT.ldr_alb_nrm.3840x2160 - CPU-Onlyhadoop: File Status - 100 - 1000000hadoop: Open - 50 - 1000000embree: Pathtracer ISPC - Asian Dragon Objhadoop: File Status - 50 - 1000000easywave: e2Asean Grid + BengkuluSept2007 Source - 1200deepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Streamdeepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Streamdeepsparse: ResNet-50, Baseline - Asynchronous Multi-Streamdeepsparse: ResNet-50, Baseline - Asynchronous Multi-Streamdeepsparse: CV Detection, YOLOv5s COCO - Asynchronous Multi-Streamdeepsparse: CV Detection, YOLOv5s COCO - Asynchronous Multi-Streamdeepsparse: CV Detection, YOLOv5s COCO, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: CV Detection, YOLOv5s COCO, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Streamdeepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Streamdeepsparse: ResNet-50, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: ResNet-50, Sparse INT8 - Asynchronous Multi-Streamsvt-av1: Preset 4 - Bosphorus 4Kembree: Pathtracer - Crownembree: Pathtracer ISPC - Crownspecfem3d: Homogeneous Halfspaceliquid-dsp: 96 - 256 - 512hadoop: Create - 100 - 1000000embree: Pathtracer - Asian Dragon Objliquid-dsp: 64 - 256 - 512liquid-dsp: 96 - 256 - 32liquid-dsp: 96 - 256 - 57embree: Pathtracer - Asian Dragonliquid-dsp: 32 - 256 - 512liquid-dsp: 64 - 256 - 32liquid-dsp: 64 - 256 - 57liquid-dsp: 16 - 256 - 512liquid-dsp: 8 - 256 - 512liquid-dsp: 4 - 256 - 512liquid-dsp: 32 - 256 - 57liquid-dsp: 32 - 256 - 32liquid-dsp: 16 - 256 - 32liquid-dsp: 1 - 256 - 512liquid-dsp: 2 - 256 - 512liquid-dsp: 16 - 256 - 57liquid-dsp: 8 - 256 - 57liquid-dsp: 8 - 256 - 32liquid-dsp: 4 - 256 - 57liquid-dsp: 4 - 256 - 32liquid-dsp: 2 - 256 - 57liquid-dsp: 2 - 256 - 32liquid-dsp: 1 - 256 - 57liquid-dsp: 1 - 256 - 32hadoop: Create - 50 - 1000000embree: Pathtracer ISPC - Asian Dragon Objembree: Pathtracer ISPC - Asian Dragonspecfem3d: Tomographic Modelspecfem3d: Mount St. Helensremhos: Sample Remap Exampleembree: Pathtracer - Crownkripke: embree: Pathtracer ISPC - Crownonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUonednn: Deconvolution Batch shapes_1d - bf16bf16bf16 - CPUonednn: Deconvolution Batch shapes_1d - f32 - CPUoidn: RT.hdr_alb_nrm.3840x2160 - CPU-Onlyoidn: RT.ldr_alb_nrm.3840x2160 - CPU-Onlyembree: Pathtracer - Asian Dragonembree: Pathtracer ISPC - Asian Dragonhadoop: Rename - 100 - 100000hadoop: Rename - 50 - 100000hadoop: Delete - 100 - 100000hadoop: Delete - 50 - 100000onednn: IP Shapes 1D - bf16bf16bf16 - CPUonednn: IP Shapes 1D - f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUhadoop: Open - 50 - 100000hadoop: Open - 100 - 100000hadoop: File Status - 100 - 100000hadoop: File Status - 50 - 100000svt-av1: Preset 4 - Bosphorus 1080phadoop: Create - 100 - 100000hadoop: Create - 50 - 100000onednn: IP Shapes 3D - bf16bf16bf16 - CPUonednn: IP Shapes 3D - f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUsvt-av1: Preset 8 - Bosphorus 4Konednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - bf16bf16bf16 - CPUsvt-av1: Preset 8 - Bosphorus 1080psvt-av1: Preset 13 - Bosphorus 4Ksvt-av1: Preset 12 - Bosphorus 4Konednn: Deconvolution Batch shapes_3d - bf16bf16bf16 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: Deconvolution Batch shapes_3d - f32 - CPUeasywave: e2Asean Grid + BengkuluSept2007 Source - 240svt-av1: Preset 12 - Bosphorus 1080psvt-av1: Preset 13 - Bosphorus 1080pabcdefg77216238331589741328790904128125401255843311592421046271275671836127087121285757254.8818095287355110580.5415.9528215.096676771000066.4224809515.986485.717549.32581110690000033.22393.630.41213.9456.0113.873914.236942.44282.5542.24283.9737.8317.2214.23842.914.882454.0916.02748.444.869837.5838.51244.690.34120606.3830.721560.034.172873.240.5486884.6416.262945.265.892033.178.285776.942.035882.9116.346826.88598380426.98502090833.3412718.918926.273078347.661268.59887323916.90741417.070690114605.038839.495198932150.5897158.924605.75839.4391118.7509201.39250.8627.35435.628672.463521533218867921126126217391374.3211322.250549.3653485.6725111.0131215.6383109.8027218.146449.0092489.12034.65085137.01145.20315.105117737116400004614553.57336225600003005800000255980000042581000022077000001994400000216080000109870000529110001192100000118350000059423000013909000279010006997400003694300003075400001962200001538500001174900007718100059401000394990005366556.485312.31294665211.02476577516.34654.901756.08711.831.8460.144967.33787552970522875669107546082942016851546452910112.477407334364990.811141.219163.013163.459422.994510.36176851736950615201312801833975925104405159728106180130802178172746489099255.31806828914243716751580.7615.9888214.074675736000066.6425666115.9785487.359949.16661124030000033.17393.2330.44213.6256.0613.766614.178342.2284.2242.09284.9937.79317.2814.03854.514.892450.2615.98750.494.859849.0738.661239.670.34120728.22311546.024.162880.580.5487359.2316.022986.465.912028.018.275780.442.055836.2716.436528.6521086329.46076119733.3798717.969326.2472129347.218968.66177167917.06691403.06586715605.730739.468197314150.6055159.0596606.669339.4539118.9482201.25280.8627.24135.6422672.37341738221619701020408194174874.5553321.182949.1082488.1264110.9152215.9254109.2332219.526348.9703489.44644.64765138.83415.14914.460586987181400004443753.81356109500002995400000257110000042962000022121000002001900000216150000108080000555880001214200000119030000060247000014021000277360006927600003669300003051100001965900001536900001140100007701900059296000394860005211956.690212.10059594711.31849570916.79155.392556.45511.831.8459.912467.19516934873046908277380146948440485845871686206912.591374254128891.322138.338166.692166.378427.686542.611762529373685963013817846939106248512681447114996217565265461189254.7223324528656540680.4115.9778214.136675417000066.7227048015.9872507.478647.15351082670000033.03393.3730.43213.7956.0213.831714.139942.43282.6742.19284.3137.79317.3314.12849.34.882455.5115.83757.384.869845.2738.751237.290.34123484.2830.891551.634.162881.140.5486789.816.022987.335.92029.798.245802.652.055840.5316.53527.49085015727.06023507933.4637716.140426.1266827347.372868.62877463816.88631418.904197031605.918339.446490147145.2562164.6131605.876539.4183118.7801201.54020.8727.40835.6825671.2554185874189393968399528425274.5031321.508249.0173489.1106111.025215.647109.5847218.516249.2055487.05224.63485153.66445.04914.8082733657150300004400153.69276226300002999800000256490000042440000022068000002010300000214910000109140000551650001254800000118480000060365000014225000282270006749300003669900003067600001945100001536700001185500007692400057519000394530005226056.932712.04091787711.3273597716.24355.403756.80781.831.8259.792967.50386715977101734759058040160640322672992765789512.617350754393790.417143.545161.495163.055431.895516.9062980641913648046977169355334487342241262258981294929814911567517612147959727670.87185632627321108224.155.57001151.9057934570000182.991978665.57469493.59616.164898.98103189000000.341643.991642.511641.92838.516847.38849.16390.03761.5910.47398.5220.035.453295.6074774.71107.0274.81106.964.41124.1221.57370.577.71036.9923.2344.674.513540.8840.4395.660.3544958.0730.02532.596.791175.670.4932002.6215.361039.6110.01797.647.932013.773.112564.786.5874571.61429432762.44174958533.2242240.55297281208325.876324.48338392115.7246508.087112613606.10113.0694111012143.764355.6132606.577313.1312112.249971.1370.3455.17331.0525257.2728124843922.35170.720.7260060127831923.3531181818238.10573.3058108.90949.0863162.8502111.108171.9189110.111472.45548.8729163.55594.9961599.20794.10721.894322.392235.5716849082862500007129622.26142829200001065200000112080000024.8452737600001059500000109330000019385000099594000502580001035000000104710000054536000012683000246270006891500003633100002780300001889300001386000001056500006705400052665000352280007213423.873327.743827.33098558826.73641411830.76121.481124099450022.5850.6282363.059913.815760.720.7224.687228.364382102823721057081010101.037492.494080.65225957803552910159171663291110.9157971586171.028751.257580.6039566.9882.133321.558241.33789118.946161.854163.1891.913740.8478053.377821.657526.216604.9862961251903678446737170853893487338813209125675976129904969071186571711724611149060145670.641855726285212714213870250224.15.56353151.5067931010000182.561957985.54107494.257516.139299.415102640000000.341643.971639.361641849.712841.078851.65990.31761.1610.47398.91205.461535.620474.5107.2774.54107.2464.68123.6121.4373.647.771028.6423.32342.814.513544.1836.98432.320.3544933.2730.1530.996.81174.60.4932032.0615.361039.8210.06793.757.962007.533.112562.546.582770.18902850662.32514682833.2625240.234971.4484360325.741624.47258404115.6245511.4098113225607.91312.9433113327144.101355.4634606.75513.1187112.057471.27270.3455.09330.9867257.894120481922.29110.720.7223562725100423.528332092438.06773.2609109.093849.0094163.1361111.088871.9146109.965472.657149.0598162.92984.98591599.15434.11421.990922.342235.0301348892858800007005722.15772818300001065100000111780000024.82822734800001057500000109540000019604000097005000503800001032000000104660000054514000012366000252070006929200003579900002777800001912300001386200001054800006884600052827000353150007089723.939327.829327.45982130826.79914344630.84521.435723624390022.56940.6339753.06373.844210.720.7224.734328.31418382282237980391006041.144322.565220.6576155248629498561349738910510.98458824586171.054251.280430.57579467.7212.12571.549111.33861119.307162.051162.6081.917810.8444343.384361.654525.173597.01129560319136125471411697549564883447032181269259541303899736811909224830148360310667.8726695210674142470105223.955.55581151.787955790000181.71962875.5732494.221116.130797.98799764500000.341642.351636.441636.76851.494845.308849.34490.26760.5710.48399.2420.015.452275.6145474.43107.3974.87106.7664.31124.321.65369.267.671041.8723.28343.494.53548.7837.01431.940.3544968.4329.95533.746.761180.850.4931951.6415.381038.4710.09791.747.972004.763.142539.976.5956370.54225590561.28176912433.2751240.164271.9685815324.956824.52388250115.721508.2088110803607.817113.0853111198143.692255.5428606.785213.0934112.41371.0440.3455.14831.0292257.5046130378122.26760.720.721964637122100123.5042179533238.01573.2163109.0949.0712162.9294110.885271.9401110.000172.573649.0714162.89764.98771600.52754.13821.774622.440735.5350730012859200007053722.1492830300001065300000112050000024.8872733900001057100000109460000019450000099441000499770001024600000104190000054502000012681000251990006933400003504500002763900001898800001385800001057400006886100052879000352710006992023.935427.82626.97375739526.87316845530.72521.591323659100022.65660.6303253.056743.818230.720.7224.704728.3237794918163399404969931.001362.497140.65318257803552356047846970922010.73659382583431.061441.206530.60083467.3932.130621.572821.34183118.486160.798161.8471.914220.8506913.379561.657521.518585.368295522191360884699317055530148934107319512627968401185491713524574148159944669.09187354169569923224.125.56539151.6817964910000183.291970925.57553495.603316.067297.529105006000000.341641.41631.991637.37848.032847.417837.59590.63759.9210.48398.1320.055.477255.6227874.71107.0474.58107.2464.77123.4121.47372.267.741031.623.42341.364.523533.6436.98432.20.3545097.9929.72538.016.791175.580.4932008.0315.371039.3710.06793.97.962006.093.122557.666.6008569.95560916562.81092437633.3674239.517672.0185763325.507524.46078481015.6892509.139113895607.162813.073110828144.105355.4264608.716313.0596112.477970.92750.3455.17231.0574257.2808110742022.25590.720.72204918065402223.7091203666037.9573.1867109.219149.0259162.9937110.975971.9003109.903772.687948.9837163.22824.97871602.52214.14321.830522.421535.3786000212865300007092222.19012817300001065700000111820000024.961927407000010562000001099300000194670000100170000495560001033400000104710000054305000012256000227270006820700003578100002774100001907500001384600001048000006867800052854000352360007270623.879627.9127.74647516227.69663137130.7521.584723717570022.77450.6291083.054583.823810.720.7224.819328.479380386822371025641039501.127232.514410.647754644846082948780556179811.01658928606801.045671.279180.6123267.8112.118131.551181.33564118.481161.324160.3221.912740.8434923.381561.648528.533586.748OpenBenchmarking.org

BRL-CAD

BRL-CAD is a cross-platform, open-source solid modeling system with built-in benchmark mode. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgVGR Performance Metric, More Is BetterBRL-CAD 7.36VGR Performance Metricgfedcba170K340K510K680K850K2955222956032961252980647625297685177721621. (CXX) g++ options: -std=c++14 -pipe -fvisibility=hidden -fno-strict-aliasing -fno-common -fexceptions -ftemplate-depth-128 -m64 -ggdb3 -O3 -fipa-pta -fstrength-reduce -finline-functions -flto -ltcl8.6 -lregex_brl -lz_brl -lnetpbm -ldl -lm -ltk8.6

OpenVKL

OpenVKL is the Intel Open Volume Kernel Library that offers high-performance volume computation kernels and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems / Sec, More Is BetterOpenVKL 2.0.0Benchmark: vklBenchmarkCPU Scalargfed4080120160200191191190191MIN: 13 / MAX: 3483MIN: 13 / MAX: 3484MIN: 13 / MAX: 3484MIN: 13 / MAX: 3471

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_read_write - Threads: 16gfedcba8K16K24K32K40K36088361253678436480373683695038331

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_read_write - Threads: 32gfedcba13K26K39K52K65K46993471414673746977596306152058974

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_non_index - Threads: 1gfedcba4008001200160020001705169717081693138113121328

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_read_write - Threads: 64gfedcba20K40K60K80K100K55301549565389355334784698018379090

OpenVKL

OpenVKL is the Intel Open Volume Kernel Library that offers high-performance volume computation kernels and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems / Sec, More Is BetterOpenVKL 2.0.0Benchmark: vklBenchmarkCPU ISPCgfed110220330440550489488487487MIN: 36 / MAX: 6969MIN: 36 / MAX: 6952MIN: 36 / MAX: 6956MIN: 36 / MAX: 6949

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_non_index - Threads: 64gfedcba9K18K27K36K45K34107344703388134224391063975941281

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_read_write - Threads: 1gfecba7001400210028003500319532183209248525102540

Test: oltp_read_write - Threads: 1

d: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_index - Threads: 16gfedca3K6K9K12K15K126271269212567126221268112558

Test: oltp_update_index - Threads: 16

b: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_point_select - Threads: 1fedcba13002600390052006500595459765898447144054331

Test: oltp_point_select - Threads: 1

g: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_point_select - Threads: 128fedcba30K60K90K120K150K130389129904129492149962159728159242

Test: oltp_point_select - Threads: 128

g: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_point_select - Threads: 32gfedba20K40K60K80K100K96840973689690798149106180104627

Test: oltp_point_select - Threads: 32

c: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_point_select - Threads: 64gfedba30K60K90K120K150K118549119092118657115675130802127567

Test: oltp_point_select - Threads: 64

c: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_index - Threads: 32gedcba4K8K12K16K20K171351711717612175651781718361

Test: oltp_update_index - Threads: 32

f: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_index - Threads: 128gfecba6K12K18K24K30K245742483024611265462746427087

Test: oltp_update_index - Threads: 128

d: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_index - Threads: 1gfedca30060090012001500148114831490147911891212

Test: oltp_update_index - Threads: 1

b: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_read_write - Threads: 128gfedba20K40K60K80K100K599446031060145597278909985757

Test: oltp_read_write - Threads: 128

c: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.6Blend File: Barbershop - Compute: CPU-Onlygfedcba140280420560700669.09667.87670.64670.87254.72255.30254.88

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_non_index - Threads: 16gedba4K8K12K16K20K1873518557185631806818095

Test: oltp_update_non_index - Threads: 16

c: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

f: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_non_index - Threads: 32fedba6K12K18K24K30K2669526285262732891428735

Test: oltp_update_non_index - Threads: 32

c: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

g: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_index - Threads: 64fedcb5K10K15K20K25K2106721271211082332424371

Test: oltp_update_index - Threads: 64

a: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

g: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_non_index - Threads: 128gfeca11K22K33K44K55K4169541424421385286551105

Test: oltp_update_non_index - Threads: 128

b: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

d: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_point_select - Threads: 16gfecb15K30K45K60K75K6992370105702506540667515

Test: oltp_point_select - Threads: 16

a: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

d: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

Timed Linux Kernel Compilation

This test times how long it takes to build the Linux kernel in a default configuration (defconfig) for the architecture being tested or alternatively an allmodconfig for building all possible kernel modules for the build. Learn more via the OpenBenchmarking.org test page.

Build: allmodconfig

a: The test quit with a non-zero exit status.

b: The test quit with a non-zero exit status.

c: The test quit with a non-zero exit status.

d: The test quit with a non-zero exit status.

e: The test quit with a non-zero exit status.

f: The test quit with a non-zero exit status.

g: The test quit with a non-zero exit status.

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.6Blend File: Pabellon Barcelona - Compute: CPU-Onlygfedcba50100150200250224.12223.95224.10224.1580.4180.7680.54

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: particle_volume/scivis/real_timegfedcba481216205.565395.555815.563535.5700115.9778015.9888015.95280

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: particle_volume/pathtracer/real_timegfedcba50100150200250151.68151.78151.51151.91214.14214.07215.10

nekRS

nekRS is an open-source Navier Stokes solver based on the spectral element method. NekRS supports both CPU and GPU/accelerator support though this test profile is currently configured for CPU execution. NekRS is part of Nek5000 of the Mathematics and Computer Science MCS at Argonne National Laboratory. This nekRS benchmark is primarily relevant to large core count HPC servers and otherwise may be very time consuming on smaller systems. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgflops/rank, More Is BetternekRS 23.0Input: TurboPipe Periodicgfedcba2000M4000M6000M8000M10000M79649100007955790000793101000079345700006754170000675736000067677100001. (CXX) g++ options: -fopenmp -O2 -march=native -mtune=native -ftree-vectorize -rdynamic -lmpi_cxx -lmpi

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.6Blend File: Classroom - Compute: CPU-Onlygfedcba4080120160200183.29181.70182.56182.9966.7266.6466.42

Apache Cassandra

This is a benchmark of the Apache Cassandra NoSQL database management system making use of cassandra-stress. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 4.1.3Test: Writesgfedcba60K120K180K240K300K197092196287195798197866270480256661248095

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: particle_volume/ao/real_timegfedcba481216205.575535.573205.541075.5746915.9872015.9785015.98600

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Streamgfedcba110220330440550495.60494.22494.26493.60507.48487.36485.72

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Streamgfedcba112233445516.0716.1316.1416.1647.1549.1749.33

easyWave

The easyWave software allows simulating tsunami generation and propagation in the context of early warning systems. EasyWave supports making use of OpenMP for CPU multi-threading and there are also GPU ports available but not currently incorporated as part of this test profile. The easyWave tsunami generation software is run with one of the example/reference input files for measuring the CPU execution time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BettereasyWave r34Input: e2Asean Grid + BengkuluSept2007 Source - Time: 2400gfed2040608010097.5397.9999.4298.981. (CXX) g++ options: -O3 -fopenmp

nekRS

nekRS is an open-source Navier Stokes solver based on the spectral element method. NekRS supports both CPU and GPU/accelerator support though this test profile is currently configured for CPU execution. NekRS is part of Nek5000 of the Mathematics and Computer Science MCS at Argonne National Laboratory. This nekRS benchmark is primarily relevant to large core count HPC servers and otherwise may be very time consuming on smaller systems. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgflops/rank, More Is BetternekRS 23.0Input: Kershawgfedcba2000M4000M6000M8000M10000M10500600000997645000010264000000103189000001082670000011240300000111069000001. (CXX) g++ options: -fopenmp -O2 -march=native -mtune=native -ftree-vectorize -rdynamic -lmpi_cxx -lmpi

Intel Open Image Denoise

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 2.1Run: RTLightmap.hdr.4096x4096 - Device: CPU-Onlygfed0.07650.1530.22950.3060.38250.340.340.340.34

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUgfed4008001200160020001641.401642.351643.971643.99MIN: 1589.91MIN: 1586.17MIN: 1590.89MIN: 1588.031. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUgfed4008001200160020001631.991636.441639.361642.51MIN: 1581.62MIN: 1585.81MIN: 1581.93MIN: 1593.161. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUgfed4008001200160020001637.371636.761641.001641.92MIN: 1584.58MIN: 1585.98MIN: 1595.55MIN: 1584.811. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUgfed2004006008001000848.03851.49849.71838.52MIN: 807.34MIN: 807.97MIN: 805.98MIN: 796.31. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUgfed2004006008001000847.42845.31841.08847.38MIN: 806.72MIN: 803.78MIN: 798.46MIN: 806.331. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUgfed2004006008001000837.60849.34851.66849.16MIN: 796.61MIN: 805.8MIN: 809.45MIN: 806.441. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.6Blend File: Fishy Cat - Compute: CPU-Onlygfedcba2040608010090.6390.2690.3190.0333.0333.1733.22

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Face Detection FP16 - Device: CPUgfedcba160320480640800759.92760.57761.16761.59393.37393.23393.60MIN: 737.63 / MAX: 771.07MIN: 741.4 / MAX: 770.88MIN: 741.99 / MAX: 776.56MIN: 738.34 / MAX: 772.36MIN: 362.57 / MAX: 433.51MIN: 360.87 / MAX: 433.13MIN: 363.29 / MAX: 431.611. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Face Detection FP16 - Device: CPUgfedcba71421283510.4810.4810.4710.4730.4330.4430.411. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Face Detection FP16-INT8 - Device: CPUgfedcba90180270360450398.13399.24398.91398.52213.79213.62213.94MIN: 379.09 / MAX: 404.71MIN: 387.9 / MAX: 408.93MIN: 386.2 / MAX: 407.29MIN: 382.1 / MAX: 404.98MIN: 197.29 / MAX: 236.32MIN: 197.2 / MAX: 235.23MIN: 201.64 / MAX: 242.711. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Face Detection FP16-INT8 - Device: CPUgfedcba132639526520.0520.0120.0020.0356.0256.0656.011. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: gravity_spheres_volume/dim_512/scivis/real_timegfedcba481216205.477255.452275.461535.4532913.8317013.7666013.87390

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: gravity_spheres_volume/dim_512/ao/real_timegfedcba481216205.622785.614545.620405.6074714.1399014.1783014.23690

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Person Detection FP16 - Device: CPUgfedcba2040608010074.7174.4374.5074.7142.4342.2042.44MIN: 66.29 / MAX: 79.68MIN: 65.68 / MAX: 83.49MIN: 66.5 / MAX: 80.32MIN: 66.12 / MAX: 81.09MIN: 36.31 / MAX: 62.36MIN: 36.84 / MAX: 61.97MIN: 36.14 / MAX: 61.981. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Person Detection FP16 - Device: CPUgfedcba60120180240300107.04107.39107.27107.02282.67284.22282.551. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Person Detection FP32 - Device: CPUgfedcba2040608010074.5874.8774.5474.8142.1942.0942.24MIN: 67.63 / MAX: 78.73MIN: 66.72 / MAX: 80.96MIN: 65.97 / MAX: 82.9MIN: 66.88 / MAX: 80.7MIN: 36.21 / MAX: 65.64MIN: 37.13 / MAX: 58.71MIN: 36.59 / MAX: 61.561. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Person Detection FP32 - Device: CPUgfedcba60120180240300107.24106.76107.24106.90284.31284.99283.971. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Machine Translation EN To DE FP16 - Device: CPUgfedcba142842567064.7764.3164.6864.4137.7937.7937.80MIN: 55.8 / MAX: 69.46MIN: 50.85 / MAX: 70.77MIN: 38.02 / MAX: 72.52MIN: 37.44 / MAX: 73.04MIN: 33.29 / MAX: 54.88MIN: 32.97 / MAX: 53.7MIN: 33.35 / MAX: 56.451. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Machine Translation EN To DE FP16 - Device: CPUgfedcba70140210280350123.41124.30123.61124.12317.33317.28317.221. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Road Segmentation ADAS FP16-INT8 - Device: CPUgfedcba51015202521.4721.6521.4021.5714.1214.0314.23MIN: 17.62 / MAX: 28.13MIN: 19.48 / MAX: 24.27MIN: 19.07 / MAX: 25.3MIN: 19.5 / MAX: 24.76MIN: 11.51 / MAX: 26.04MIN: 11.59 / MAX: 26.04MIN: 11.51 / MAX: 25.861. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Road Segmentation ADAS FP16-INT8 - Device: CPUgfedcba2004006008001000372.26369.26373.64370.57849.30854.51842.911. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Person Vehicle Bike Detection FP16 - Device: CPUgfedcba2468107.747.677.777.704.884.894.88MIN: 6.06 / MAX: 12.66MIN: 5.32 / MAX: 16.6MIN: 5.42 / MAX: 16.35MIN: 5.51 / MAX: 16.06MIN: 3.9 / MAX: 14.94MIN: 3.93 / MAX: 13.44MIN: 3.95 / MAX: 16.051. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Person Vehicle Bike Detection FP16 - Device: CPUgfedcba50010001500200025001031.601041.871028.641036.992455.512450.262454.091. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Road Segmentation ADAS FP16 - Device: CPUgfedcba61218243023.4223.2823.3223.2015.8315.9816.02MIN: 20.46 / MAX: 32.43MIN: 15.73 / MAX: 30.77MIN: 19.49 / MAX: 30.99MIN: 15.1 / MAX: 31.6MIN: 12.38 / MAX: 32.97MIN: 12.74 / MAX: 33.34MIN: 12.5 / MAX: 33.941. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Road Segmentation ADAS FP16 - Device: CPUgfedcba160320480640800341.36343.49342.81344.67757.38750.49748.441. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Face Detection Retail FP16-INT8 - Device: CPUgfedcba1.09352.1873.28054.3745.46754.524.504.514.514.864.854.86MIN: 2.77 / MAX: 13.57MIN: 2.98 / MAX: 13.86MIN: 2.96 / MAX: 16.06MIN: 2.98 / MAX: 13.05MIN: 4.34 / MAX: 12.27MIN: 4.25 / MAX: 12.86MIN: 4.23 / MAX: 12.811. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Face Detection Retail FP16-INT8 - Device: CPUgfedcba2K4K6K8K10K3533.643548.783544.183540.889845.279849.079837.581. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Handwritten English Recognition FP16-INT8 - Device: CPUgfedcba91827364536.9837.0136.9840.4038.7538.6638.50MIN: 32.61 / MAX: 41.91MIN: 32.25 / MAX: 43.6MIN: 32.02 / MAX: 44.78MIN: 26.93 / MAX: 74.83MIN: 37.46 / MAX: 43.52MIN: 37.22 / MAX: 43.52MIN: 36.77 / MAX: 44.231. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Handwritten English Recognition FP16-INT8 - Device: CPUgfedcba30060090012001500432.20431.94432.32395.661237.291239.671244.691. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUgfedcba0.07880.15760.23640.31520.3940.350.350.350.350.340.340.34MIN: 0.23 / MAX: 8.63MIN: 0.23 / MAX: 9.15MIN: 0.23 / MAX: 8.84MIN: 0.23 / MAX: 9.09MIN: 0.29 / MAX: 7.09MIN: 0.29 / MAX: 10.87MIN: 0.29 / MAX: 7.331. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUgfedcba30K60K90K120K150K45097.9944968.4344933.2744958.07123484.28120728.22120606.381. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Handwritten English Recognition FP16 - Device: CPUgfedcba71421283529.7229.9530.1030.0230.8931.0030.72MIN: 19.46 / MAX: 38.99MIN: 19.01 / MAX: 38.08MIN: 22.61 / MAX: 39.15MIN: 18.78 / MAX: 38.72MIN: 29.48 / MAX: 36.29MIN: 29.59 / MAX: 36.33MIN: 29.51 / MAX: 35.071. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Handwritten English Recognition FP16 - Device: CPUgfedcba30060090012001500538.01533.74530.99532.591551.631546.021560.031. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Vehicle Detection FP16-INT8 - Device: CPUgfedcba2468106.796.766.806.794.164.164.17MIN: 3.79 / MAX: 15.41MIN: 4.04 / MAX: 15.47MIN: 4.04 / MAX: 15.37MIN: 3.8 / MAX: 15.48MIN: 3.43 / MAX: 10.26MIN: 3.42 / MAX: 11.2MIN: 3.39 / MAX: 10.071. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Vehicle Detection FP16-INT8 - Device: CPUgfedcba60012001800240030001175.581180.851174.601175.672881.142880.582873.241. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Age Gender Recognition Retail 0013 FP16 - Device: CPUgfedcba0.12150.2430.36450.4860.60750.490.490.490.490.540.540.54MIN: 0.3 / MAX: 8.84MIN: 0.3 / MAX: 8.2MIN: 0.3 / MAX: 9.07MIN: 0.3 / MAX: 9.28MIN: 0.45 / MAX: 5.03MIN: 0.45 / MAX: 7.81MIN: 0.45 / MAX: 7.641. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Age Gender Recognition Retail 0013 FP16 - Device: CPUgfedcba20K40K60K80K100K32008.0331951.6432032.0632002.6286789.8087359.2386884.641. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Weld Porosity Detection FP16 - Device: CPUgfedcba4812162015.3715.3815.3615.3616.0216.0216.26MIN: 7.99 / MAX: 23.98MIN: 7.99 / MAX: 24MIN: 8.02 / MAX: 23.81MIN: 8.08 / MAX: 24.34MIN: 14.63 / MAX: 33.79MIN: 14.41 / MAX: 30.55MIN: 14.71 / MAX: 28.141. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Weld Porosity Detection FP16 - Device: CPUgfedcba60012001800240030001039.371038.471039.821039.612987.332986.462945.261. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Vehicle Detection FP16 - Device: CPUgfedcba369121510.0610.0910.0610.015.905.915.89MIN: 5.2 / MAX: 19.38MIN: 5.4 / MAX: 19.17MIN: 5.29 / MAX: 19.07MIN: 5.7 / MAX: 19.52MIN: 4.83 / MAX: 13.4MIN: 4.84 / MAX: 12.9MIN: 4.67 / MAX: 18.41. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Vehicle Detection FP16 - Device: CPUgfedcba400800120016002000793.90791.74793.75797.642029.792028.012033.171. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Weld Porosity Detection FP16-INT8 - Device: CPUgfedcba2468107.967.977.967.938.248.278.28MIN: 4.19 / MAX: 14.2MIN: 4.37 / MAX: 16.86MIN: 4.19 / MAX: 16.59MIN: 4.2 / MAX: 16.92MIN: 7.62 / MAX: 23.32MIN: 7.37 / MAX: 25.18MIN: 7.44 / MAX: 23.351. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Weld Porosity Detection FP16-INT8 - Device: CPUgfedcba120024003600480060002006.092004.762007.532013.775802.655780.445776.941. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Face Detection Retail FP16 - Device: CPUgfedcba0.70651.4132.11952.8263.53253.123.143.113.112.052.052.03MIN: 1.88 / MAX: 11.92MIN: 1.93 / MAX: 11.65MIN: 1.93 / MAX: 9.72MIN: 1.94 / MAX: 11.57MIN: 1.62 / MAX: 6.96MIN: 1.6 / MAX: 7MIN: 1.66 / MAX: 7.511. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Face Detection Retail FP16 - Device: CPUgfedcba130026003900520065002557.662539.972562.542564.785840.535836.275882.911. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_timegfedcba481216206.600856.595636.582706.5874516.5350016.4365016.34680

SPECFEM3D

simulates acoustic (fluid), elastic (solid), coupled acoustic/elastic, poroelastic or seismic wave propagation in any type of conforming mesh of hexahedra. This test profile currently relies on CPU-based execution for SPECFEM3D and using a variety of their built-in examples/models for benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSPECFEM3D 4.0Model: Layered Halfspacegfedcba163248648069.9670.5470.1971.6127.4928.6526.891. (F9X) gfortran options: -O2 -fopenmp -std=f2003 -fimplicit-none -fmax-errors=10 -pedantic -pedantic-errors -O3 -finline-functions -lmpi_usempif08 -lmpi_mpifh -lmpi

OpenBenchmarking.orgSeconds, Fewer Is BetterSPECFEM3D 4.0Model: Water-layered Halfspacegfedcba142842567062.8161.2862.3362.4427.0629.4626.991. (F9X) gfortran options: -O2 -fopenmp -std=f2003 -fimplicit-none -fmax-errors=10 -pedantic -pedantic-errors -O3 -finline-functions -lmpi_usempif08 -lmpi_mpifh -lmpi

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Streamgfedcba81624324033.3733.2833.2633.2233.4633.3833.34

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Streamgfedcba160320480640800239.52240.16240.23240.55716.14717.97718.92

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.6Blend File: BMW27 - Compute: CPU-Onlygfedcba163248648072.0171.9671.4472.0026.1226.2426.20

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Rename - Threads: 100 - Files: 1000000gfedcba20K40K60K80K100K85763858158436081208668277212973078

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Streamgfedcba80160240320400325.51324.96325.74325.88347.37347.22347.66

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Streamgfedcba153045607524.4624.5224.4724.4868.6368.6668.60

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Rename - Threads: 50 - Files: 1000000gfedcba20K40K60K80K100K84810825018404183921746387167973239

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Streamgfedcba4812162015.6915.7215.6215.7216.8917.0716.91

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Streamgfedcba30060090012001500509.14508.21511.41508.091418.901403.071417.07

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Delete - Threads: 100 - Files: 1000000gfedcba20K40K60K80K100K113895110803113225112613970318671590114

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Streamgfedcba130260390520650607.16607.82607.91606.10605.92605.73605.04

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Streamgfedcba91827364513.0713.0912.9413.0739.4539.4739.50

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Delete - Threads: 50 - Files: 1000000gfedcba20K40K60K80K100K110828111198113327111012901479731498932

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-Streamgfedcba306090120150144.11143.69144.10143.76145.26150.61150.59

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-Streamgfedcba408012016020055.4355.5455.4655.61164.61159.06158.92

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Streamgfedcba130260390520650608.72606.79606.76606.58605.88606.67605.76

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Streamgfedcba91827364513.0613.0913.1213.1339.4239.4539.44

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-Streamgfedcba306090120150112.48112.41112.06112.25118.78118.95118.75

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-Streamgfedcba408012016020070.9371.0471.2771.14201.54201.25201.39

Intel Open Image Denoise

Open Image Denoise is a denoising library for ray-tracing and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 2.0Run: RTLightmap.hdr.4096x4096 - Device: CPU-Onlygfedcba0.19580.39160.58740.78320.9790.340.340.340.340.870.860.86

Timed Linux Kernel Compilation

This test times how long it takes to build the Linux kernel in a default configuration (defconfig) for the architecture being tested or alternatively an allmodconfig for building all possible kernel modules for the build. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 6.1Build: defconfiggfedcba122436486055.1755.1555.0955.1727.4127.2427.35

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Asynchronous Multi-Streamgfedcba81624324031.0631.0330.9931.0535.6835.6435.63

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Asynchronous Multi-Streamgfedcba150300450600750257.28257.50257.89257.27671.26672.37672.46

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Open - Threads: 100 - Files: 1000000gfedcba300K600K900K1200K1500K1107420130378112048191248439185874173822215332

Embree

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer - Model: Asian Dragon Objgfed51015202522.2622.2722.2922.35MIN: 22.18 / MAX: 22.43MIN: 22.2 / MAX: 22.44MIN: 22.22 / MAX: 22.46MIN: 22.28 / MAX: 22.5

Intel Open Image Denoise

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 2.1Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Onlygfed0.1620.3240.4860.6480.810.720.720.720.72

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 2.1Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Onlygfed0.1620.3240.4860.6480.810.720.720.720.72

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: File Status - Threads: 100 - Files: 1000000gfedcba400K800K1200K1600K2000K2049180196463723562760060118939391619701886792

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Open - Threads: 50 - Files: 1000000gfedcba300K600K900K1200K1500K654022122100125100427831968399510204081126126

Embree

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer ISPC - Model: Asian Dragon Objgfed61218243023.7123.5023.5323.35MIN: 23.61 / MAX: 23.93MIN: 23.4 / MAX: 23.74MIN: 23.43 / MAX: 23.73MIN: 23.26 / MAX: 23.57

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: File Status - Threads: 50 - Files: 1000000gfedcba500K1000K1500K2000K2500K20366601795332320924181818228425219417482173913

easyWave

The easyWave software allows simulating tsunami generation and propagation in the context of early warning systems. EasyWave supports making use of OpenMP for CPU multi-threading and there are also GPU ports available but not currently incorporated as part of this test profile. The easyWave tsunami generation software is run with one of the example/reference input files for measuring the CPU execution time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BettereasyWave r34Input: e2Asean Grid + BengkuluSept2007 Source - Time: 1200gfed91827364537.9538.0238.0738.111. (CXX) g++ options: -O3 -fopenmp

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Streamgfedcba2040608010073.1973.2273.2673.3174.5074.5674.32

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Streamgfedcba70140210280350109.22109.09109.09108.91321.51321.18322.25

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Streamgfedcba112233445549.0349.0749.0149.0949.0249.1149.37

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Streamgfedcba110220330440550162.99162.93163.14162.85489.11488.13485.67

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Streamgfedcba20406080100110.98110.89111.09111.11111.03110.92111.01

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Streamgfedcba5010015020025071.9071.9471.9171.92215.65215.93215.64

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Streamgfedcba20406080100109.90110.00109.97110.11109.58109.23109.80

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Streamgfedcba5010015020025072.6972.5772.6672.46218.52219.53218.15

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Streamgfedcba112233445548.9849.0749.0648.8749.2148.9749.01

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Streamgfedcba110220330440550163.23162.90162.93163.56487.05489.45489.12

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Streamgfedcba1.12412.24823.37234.49645.62054.97874.98774.98594.99604.63484.64764.6508

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Streamgfedcba110022003300440055001602.521600.531599.151599.215153.665138.835137.01

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 4 - Input: Bosphorus 4Kgfedcba1.17072.34143.51214.68285.85354.1434.1384.1144.1075.0495.1495.2031. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

Embree

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer - Model: Crowngfed51015202521.8321.7721.9921.89MIN: 21.69 / MAX: 22.17MIN: 21.63 / MAX: 22.18MIN: 21.84 / MAX: 22.32MIN: 21.74 / MAX: 22.23

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer ISPC - Model: Crowngfed51015202522.4222.4422.3422.39MIN: 22.22 / MAX: 22.85MIN: 22.25 / MAX: 22.78MIN: 22.15 / MAX: 22.75MIN: 22.2 / MAX: 22.85

SPECFEM3D

simulates acoustic (fluid), elastic (solid), coupled acoustic/elastic, poroelastic or seismic wave propagation in any type of conforming mesh of hexahedra. This test profile currently relies on CPU-based execution for SPECFEM3D and using a variety of their built-in examples/models for benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSPECFEM3D 4.0Model: Homogeneous Halfspacegfedcba81624324035.3835.5435.0335.5714.8114.4615.111. (F9X) gfortran options: -O2 -fopenmp -std=f2003 -fimplicit-none -fmax-errors=10 -pedantic -pedantic-errors -O3 -finline-functions -lmpi_usempif08 -lmpi_mpifh -lmpi

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 96 - Buffer Length: 256 - Filter Length: 512gfedcba150M300M450M600M750M2865300002859200002858800002862500007150300007181400007116400001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Create - Threads: 100 - Files: 1000000gfedcba15K30K45K60K75K70922705377005771296440014443746145

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer - Model: Asian Dragon Objgfedcba122436486022.1922.1522.1622.2653.6953.8153.57MIN: 22.12 / MAX: 22.33MIN: 22.07 / MAX: 22.32MIN: 22.08 / MAX: 22.35MIN: 22.18 / MAX: 22.42MIN: 52.63 / MAX: 55.24MIN: 52.72 / MAX: 55.86MIN: 52.17 / MAX: 55.38

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 64 - Buffer Length: 256 - Filter Length: 512gfedcba130M260M390M520M650M2817300002830300002818300002829200006226300006109500006225600001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 96 - Buffer Length: 256 - Filter Length: 32gfedcba600M1200M1800M2400M3000M10657000001065300000106510000010652000002999800000299540000030058000001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 96 - Buffer Length: 256 - Filter Length: 57gfedcba600M1200M1800M2400M3000M11182000001120500000111780000011208000002564900000257110000025598000001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

Embree

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer - Model: Asian Dragongfed61218243024.9624.8924.8324.85MIN: 24.9 / MAX: 25.13MIN: 24.81 / MAX: 25.06MIN: 24.76 / MAX: 24.96MIN: 24.78 / MAX: 25

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 32 - Buffer Length: 256 - Filter Length: 512gfedcba90M180M270M360M450M2740700002733900002734800002737600004244000004296200004258100001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 64 - Buffer Length: 256 - Filter Length: 32gfedcba500M1000M1500M2000M2500M10562000001057100000105750000010595000002206800000221210000022077000001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 64 - Buffer Length: 256 - Filter Length: 57gfedcba400M800M1200M1600M2000M10993000001094600000109540000010933000002010300000200190000019944000001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 16 - Buffer Length: 256 - Filter Length: 512gfedcba50M100M150M200M250M1946700001945000001960400001938500002149100002161500002160800001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 8 - Buffer Length: 256 - Filter Length: 512gfedcba20M40M60M80M100M1001700009944100097005000995940001091400001080800001098700001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 4 - Buffer Length: 256 - Filter Length: 512gfedcba12M24M36M48M60M495560004997700050380000502580005516500055588000529110001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 32 - Buffer Length: 256 - Filter Length: 57gfedcba300M600M900M1200M1500M10334000001024600000103200000010350000001254800000121420000011921000001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 32 - Buffer Length: 256 - Filter Length: 32gfedcba300M600M900M1200M1500M10471000001041900000104660000010471000001184800000119030000011835000001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 16 - Buffer Length: 256 - Filter Length: 32gfedcba130M260M390M520M650M5430500005450200005451400005453600006036500006024700005942300001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 1 - Buffer Length: 256 - Filter Length: 512gfedcba3M6M9M12M15M122560001268100012366000126830001422500014021000139090001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 2 - Buffer Length: 256 - Filter Length: 512gfedcba6M12M18M24M30M227270002519900025207000246270002822700027736000279010001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 16 - Buffer Length: 256 - Filter Length: 57gfedcba150M300M450M600M750M6820700006933400006929200006891500006749300006927600006997400001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 8 - Buffer Length: 256 - Filter Length: 57gfedcba80M160M240M320M400M3578100003504500003579900003633100003669900003669300003694300001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 8 - Buffer Length: 256 - Filter Length: 32gfedcba70M140M210M280M350M2774100002763900002777800002780300003067600003051100003075400001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 4 - Buffer Length: 256 - Filter Length: 57gfedcba40M80M120M160M200M1907500001898800001912300001889300001945100001965900001962200001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 4 - Buffer Length: 256 - Filter Length: 32gfedcba30M60M90M120M150M1384600001385800001386200001386000001536700001536900001538500001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 2 - Buffer Length: 256 - Filter Length: 57gfedcba30M60M90M120M150M1048000001057400001054800001056500001185500001140100001174900001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 2 - Buffer Length: 256 - Filter Length: 32gfedcba17M34M51M68M85M686780006886100068846000670540007692400077019000771810001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 1 - Buffer Length: 256 - Filter Length: 57gfedcba13M26M39M52M65M528540005287900052827000526650005751900059296000594010001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 1 - Buffer Length: 256 - Filter Length: 32gfedcba8M16M24M32M40M352360003527100035315000352280003945300039486000394990001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Create - Threads: 50 - Files: 1000000gfedcba16K32K48K64K80K72706699207089772134522605211953665

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer ISPC - Model: Asian Dragon Objgfedcba132639526523.8823.9423.9423.8756.9356.6956.49MIN: 23.79 / MAX: 24.08MIN: 23.84 / MAX: 24.16MIN: 23.84 / MAX: 24.18MIN: 23.78 / MAX: 24.08MIN: 55.56 / MAX: 59.67MIN: 55.42 / MAX: 58.97MIN: 55.29 / MAX: 58.38

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer ISPC - Model: Asian Dragongfed71421283527.9127.8327.8327.74MIN: 27.81 / MAX: 28.17MIN: 27.73 / MAX: 28.13MIN: 27.72 / MAX: 28.1MIN: 27.64 / MAX: 27.98

SPECFEM3D

simulates acoustic (fluid), elastic (solid), coupled acoustic/elastic, poroelastic or seismic wave propagation in any type of conforming mesh of hexahedra. This test profile currently relies on CPU-based execution for SPECFEM3D and using a variety of their built-in examples/models for benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSPECFEM3D 4.0Model: Tomographic Modelgfedcba71421283527.7526.9727.4627.3312.0412.1012.311. (F9X) gfortran options: -O2 -fopenmp -std=f2003 -fimplicit-none -fmax-errors=10 -pedantic -pedantic-errors -O3 -finline-functions -lmpi_usempif08 -lmpi_mpifh -lmpi

OpenBenchmarking.orgSeconds, Fewer Is BetterSPECFEM3D 4.0Model: Mount St. Helensgfedcba71421283527.7026.8726.8026.7411.3311.3211.021. (F9X) gfortran options: -O2 -fopenmp -std=f2003 -fimplicit-none -fmax-errors=10 -pedantic -pedantic-errors -O3 -finline-functions -lmpi_usempif08 -lmpi_mpifh -lmpi

Remhos

Remhos (REMap High-Order Solver) is a miniapp that solves the pure advection equations that are used to perform monotonic and conservative discontinuous field interpolation (remap) as part of the Eulerian phase in Arbitrary Lagrangian Eulerian (ALE) simulations. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterRemhos 1.0Test: Sample Remap Examplegfedcba71421283530.7530.7330.8530.7616.2416.7916.351. (CXX) g++ options: -O3 -std=c++11 -lmfem -lHYPRE -lmetis -lrt -lmpi_cxx -lmpi

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/ and https://github.com/OpenRadioss/ModelExchange/tree/main/Examples. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

Model: Chrysler Neon 1M

a: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

b: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

c: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

d: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

e: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

f: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

g: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer - Model: Crowngfedcba122436486021.5821.5921.4421.4855.4055.3954.90MIN: 21.43 / MAX: 21.89MIN: 21.45 / MAX: 21.84MIN: 21.3 / MAX: 21.78MIN: 21.32 / MAX: 21.8MIN: 53.71 / MAX: 58.99MIN: 54.02 / MAX: 57.64MIN: 53.27 / MAX: 57.28

Kripke

Kripke is a simple, scalable, 3D Sn deterministic particle transport code. Its primary purpose is to research how data layout, programming paradigms and architectures effect the implementation and performance of Sn transport. Kripke is developed by LLNL. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgThroughput FoM, More Is BetterKripke 1.2.6gfed50M100M150M200M250M2371757002365910002362439002409945001. (CXX) g++ options: -O3 -fopenmp -ldl

a: The test quit with a non-zero exit status.

b: The test quit with a non-zero exit status.

c: The test quit with a non-zero exit status.

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer ISPC - Model: Crowngfedcba132639526522.7722.6622.5722.5956.8156.4656.09MIN: 22.57 / MAX: 23.16MIN: 22.45 / MAX: 22.99MIN: 22.39 / MAX: 22.93MIN: 22.39 / MAX: 22.98MIN: 55.27 / MAX: 59.91MIN: 54.53 / MAX: 59.89MIN: 54.05 / MAX: 59.82

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUgfed0.14260.28520.42780.57040.7130.6291080.6303250.6339750.628236MIN: 0.6MIN: 0.6MIN: 0.6MIN: 0.61. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPUgfed0.68931.37862.06792.75723.44653.054583.056743.063703.05991MIN: 2.97MIN: 2.97MIN: 2.97MIN: 2.961. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPUgfed0.86491.72982.59473.45964.32453.823813.818233.844213.81576MIN: 3.29MIN: 3.25MIN: 3.27MIN: 3.261. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Intel Open Image Denoise

Open Image Denoise is a denoising library for ray-tracing and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 2.0Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Onlygfedcba0.41180.82361.23541.64722.0590.720.720.720.721.831.831.83

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 2.0Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Onlygfedcba0.4140.8281.2421.6562.070.720.720.720.721.821.841.84

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer - Model: Asian Dragongfedcba132639526524.8224.7024.7324.6959.7959.9160.14MIN: 24.74 / MAX: 25MIN: 24.63 / MAX: 24.84MIN: 24.67 / MAX: 24.86MIN: 24.62 / MAX: 24.84MIN: 58.46 / MAX: 62.03MIN: 58.66 / MAX: 61.96MIN: 58.97 / MAX: 62

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer ISPC - Model: Asian Dragongfedcba153045607528.4828.3228.3128.3667.5067.2067.34MIN: 28.37 / MAX: 28.69MIN: 28.23 / MAX: 28.55MIN: 28.21 / MAX: 28.56MIN: 28.26 / MAX: 28.59MIN: 65.64 / MAX: 71.17MIN: 65.48 / MAX: 70.41MIN: 65.61 / MAX: 70.54

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Rename - Threads: 100 - Files: 100000gfedcba20K40K60K80K100K80386794918382282102671596934875529

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Rename - Threads: 50 - Files: 100000gfedcba20K40K60K80K100K82237816338223782372771017304670522

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Delete - Threads: 100 - Files: 100000gfedcba20K40K60K80K100K1025649940498039105708734759082787566

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Delete - Threads: 50 - Files: 100000gfedcba20K40K60K80K100K10395096993100604101010905807380191075

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPUgfed0.25750.5150.77251.031.28751.127231.001361.144321.03749MIN: 0.93MIN: 0.92MIN: 1.07MIN: 0.921. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUgfed0.57721.15441.73162.30882.8862.514412.497142.565222.49408MIN: 2.3MIN: 2.26MIN: 2.32MIN: 2.31. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUgfed0.1480.2960.4440.5920.740.6477000.6531820.6576100.652259MIN: 0.57MIN: 0.57MIN: 0.57MIN: 0.571. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Open - Threads: 50 - Files: 100000gfedcba120K240K360K480K600K546448578035552486578035401606469484460829

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Open - Threads: 100 - Files: 100000gfedcba110K220K330K440K550K460829523560294985529101403226404858420168

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: File Status - Threads: 100 - Files: 100000gfedcba160K320K480K640K800K487805478469613497591716729927458716515464

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: File Status - Threads: 50 - Files: 100000gfedcba200K400K600K800K1000K561798709220389105632911657895862069529101

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 4 - Input: Bosphorus 1080pgfedcba369121511.0210.7410.9810.9112.6212.5912.481. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Create - Threads: 100 - Files: 100000gfedcba13K26K39K52K65K58928593825882457971350753742540733

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Create - Threads: 50 - Files: 100000gfedcba13K26K39K52K65K60680583435861758617439374128843649

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPUgfed0.23880.47760.71640.95521.1941.045671.061441.054251.02875MIN: 0.98MIN: 0.98MIN: 0.97MIN: 0.961. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUgfed0.28810.57620.86431.15241.44051.279181.206531.280431.25758MIN: 1.24MIN: 1.18MIN: 1.24MIN: 1.211. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUgfed0.13780.27560.41340.55120.6890.6123200.6008340.5757940.603950MIN: 0.53MIN: 0.53MIN: 0.52MIN: 0.531. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 8 - Input: Bosphorus 4Kgfedcba2040608010067.8167.3967.7266.9990.4291.3290.811. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/ and https://github.com/OpenRadioss/ModelExchange/tree/main/Examples. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

Model: Cell Phone Drop Test

a: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

b: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

c: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

d: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

e: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

f: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

g: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUgfed0.480.961.441.922.42.118132.130622.125702.13332MIN: 1.99MIN: 1.97MIN: 2.01MIN: 21. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUgfed0.35390.70781.06171.41561.76951.551181.572821.549111.55824MIN: 1.52MIN: 1.53MIN: 1.51MIN: 1.511. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPUgfed0.30190.60380.90571.20761.50951.335641.341831.338611.33789MIN: 1.31MIN: 1.31MIN: 1.31MIN: 1.311. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 8 - Input: Bosphorus 1080pgfedcba306090120150118.48118.49119.31118.95143.55138.34141.221. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 13 - Input: Bosphorus 4Kgfedcba4080120160200161.32160.80162.05161.85161.50166.69163.011. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 12 - Input: Bosphorus 4Kgfedcba4080120160200160.32161.85162.61163.19163.06166.38163.461. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/ and https://github.com/OpenRadioss/ModelExchange/tree/main/Examples. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

Model: Rubber O-Ring Seal Installation

a: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

b: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

c: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

d: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

e: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

f: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

g: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

Model: Bird Strike on Windshield

a: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

b: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

c: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

d: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

e: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

f: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

g: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

Model: Bumper Beam

a: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

b: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

c: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

d: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

e: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

f: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

g: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPUgfed0.43150.8631.29451.7262.15751.912741.914221.917811.91374MIN: 1.88MIN: 1.88MIN: 1.88MIN: 1.881. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUgfed0.19140.38280.57420.76560.9570.8434920.8506910.8444340.847805MIN: 0.83MIN: 0.83MIN: 0.83MIN: 0.831. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPUgfed0.76151.5232.28453.0463.80753.381563.379563.384363.37782MIN: 3.33MIN: 3.33MIN: 3.33MIN: 3.331. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

easyWave

The easyWave software allows simulating tsunami generation and propagation in the context of early warning systems. EasyWave supports making use of OpenMP for CPU multi-threading and there are also GPU ports available but not currently incorporated as part of this test profile. The easyWave tsunami generation software is run with one of the example/reference input files for measuring the CPU execution time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BettereasyWave r34Input: e2Asean Grid + BengkuluSept2007 Source - Time: 240gfed0.37280.74561.11841.49121.8641.6481.6571.6541.6571. (CXX) g++ options: -O3 -fopenmp

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/ and https://github.com/OpenRadioss/ModelExchange/tree/main/Examples. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

Model: INIVOL and Fluid Structure Interaction Drop Container

a: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

b: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

c: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

d: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

e: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

f: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

g: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 12 - Input: Bosphorus 1080pgfedcba110220330440550528.53521.52525.17526.22431.90427.69422.991. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 13 - Input: Bosphorus 1080pgfedcba130260390520650586.75585.37597.01604.99516.91542.61510.361. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

205 Results Shown

BRL-CAD
OpenVKL
TiDB Community Server:
  oltp_read_write - 16
  oltp_read_write - 32
  oltp_update_non_index - 1
  oltp_read_write - 64
OpenVKL
TiDB Community Server:
  oltp_update_non_index - 64
  oltp_read_write - 1
  oltp_update_index - 16
  oltp_point_select - 1
  oltp_point_select - 128
  oltp_point_select - 32
  oltp_point_select - 64
  oltp_update_index - 32
  oltp_update_index - 128
  oltp_update_index - 1
  oltp_read_write - 128
Blender
TiDB Community Server:
  oltp_update_non_index - 16
  oltp_update_non_index - 32
  oltp_update_index - 64
  oltp_update_non_index - 128
  oltp_point_select - 16
Blender
OSPRay:
  particle_volume/scivis/real_time
  particle_volume/pathtracer/real_time
nekRS
Blender
Apache Cassandra
OSPRay
Neural Magic DeepSparse:
  BERT-Large, NLP Question Answering - Asynchronous Multi-Stream:
    ms/batch
    items/sec
easyWave
nekRS
Intel Open Image Denoise
oneDNN:
  Recurrent Neural Network Training - bf16bf16bf16 - CPU
  Recurrent Neural Network Training - u8s8f32 - CPU
  Recurrent Neural Network Training - f32 - CPU
  Recurrent Neural Network Inference - f32 - CPU
  Recurrent Neural Network Inference - bf16bf16bf16 - CPU
  Recurrent Neural Network Inference - u8s8f32 - CPU
Blender
OpenVINO:
  Face Detection FP16 - CPU:
    ms
    FPS
  Face Detection FP16-INT8 - CPU:
    ms
    FPS
OSPRay:
  gravity_spheres_volume/dim_512/scivis/real_time
  gravity_spheres_volume/dim_512/ao/real_time
OpenVINO:
  Person Detection FP16 - CPU:
    ms
    FPS
  Person Detection FP32 - CPU:
    ms
    FPS
  Machine Translation EN To DE FP16 - CPU:
    ms
    FPS
  Road Segmentation ADAS FP16-INT8 - CPU:
    ms
    FPS
  Person Vehicle Bike Detection FP16 - CPU:
    ms
    FPS
  Road Segmentation ADAS FP16 - CPU:
    ms
    FPS
  Face Detection Retail FP16-INT8 - CPU:
    ms
    FPS
  Handwritten English Recognition FP16-INT8 - CPU:
    ms
    FPS
  Age Gender Recognition Retail 0013 FP16-INT8 - CPU:
    ms
    FPS
  Handwritten English Recognition FP16 - CPU:
    ms
    FPS
  Vehicle Detection FP16-INT8 - CPU:
    ms
    FPS
  Age Gender Recognition Retail 0013 FP16 - CPU:
    ms
    FPS
  Weld Porosity Detection FP16 - CPU:
    ms
    FPS
  Vehicle Detection FP16 - CPU:
    ms
    FPS
  Weld Porosity Detection FP16-INT8 - CPU:
    ms
    FPS
  Face Detection Retail FP16 - CPU:
    ms
    FPS
OSPRay
SPECFEM3D:
  Layered Halfspace
  Water-layered Halfspace
Neural Magic DeepSparse:
  BERT-Large, NLP Question Answering, Sparse INT8 - Asynchronous Multi-Stream:
    ms/batch
    items/sec
Blender
Apache Hadoop
Neural Magic DeepSparse:
  CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Stream:
    ms/batch
    items/sec
Apache Hadoop
Neural Magic DeepSparse:
  NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Asynchronous Multi-Stream:
    ms/batch
    items/sec
Apache Hadoop
Neural Magic DeepSparse:
  NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Stream:
    ms/batch
    items/sec
Apache Hadoop
Neural Magic DeepSparse:
  NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Asynchronous Multi-Stream:
    ms/batch
    items/sec
Intel Open Image Denoise
Timed Linux Kernel Compilation
Neural Magic DeepSparse:
  NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Asynchronous Multi-Stream:
    ms/batch
    items/sec
Apache Hadoop
Embree
Intel Open Image Denoise:
  RT.hdr_alb_nrm.3840x2160 - CPU-Only
  RT.ldr_alb_nrm.3840x2160 - CPU-Only
Apache Hadoop:
  File Status - 100 - 1000000
  Open - 50 - 1000000
Embree
Apache Hadoop
easyWave
Neural Magic DeepSparse:
  NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  ResNet-50, Baseline - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  CV Detection, YOLOv5s COCO - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  CV Detection, YOLOv5s COCO, Sparse INT8 - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Stream:
    ms/batch
    items/sec
  ResNet-50, Sparse INT8 - Asynchronous Multi-Stream:
    ms/batch
    items/sec
SVT-AV1
Embree:
  Pathtracer - Crown
  Pathtracer ISPC - Crown
SPECFEM3D
Liquid-DSP
Apache Hadoop
Embree
Liquid-DSP:
  64 - 256 - 512
  96 - 256 - 32
  96 - 256 - 57
Embree
Liquid-DSP:
  32 - 256 - 512
  64 - 256 - 32
  64 - 256 - 57
  16 - 256 - 512
  8 - 256 - 512
  4 - 256 - 512
  32 - 256 - 57
  32 - 256 - 32
  16 - 256 - 32
  1 - 256 - 512
  2 - 256 - 512
  16 - 256 - 57
  8 - 256 - 57
  8 - 256 - 32
  4 - 256 - 57
  4 - 256 - 32
  2 - 256 - 57
  2 - 256 - 32
  1 - 256 - 57
  1 - 256 - 32
Apache Hadoop
Embree
Embree
SPECFEM3D:
  Tomographic Model
  Mount St. Helens
Remhos
Embree
Kripke
Embree
oneDNN:
  Deconvolution Batch shapes_1d - u8s8f32 - CPU
  Deconvolution Batch shapes_1d - bf16bf16bf16 - CPU
  Deconvolution Batch shapes_1d - f32 - CPU
Intel Open Image Denoise:
  RT.hdr_alb_nrm.3840x2160 - CPU-Only
  RT.ldr_alb_nrm.3840x2160 - CPU-Only
Embree:
  Pathtracer - Asian Dragon
  Pathtracer ISPC - Asian Dragon
Apache Hadoop:
  Rename - 100 - 100000
  Rename - 50 - 100000
  Delete - 100 - 100000
  Delete - 50 - 100000
oneDNN:
  IP Shapes 1D - bf16bf16bf16 - CPU
  IP Shapes 1D - f32 - CPU
  IP Shapes 1D - u8s8f32 - CPU
Apache Hadoop:
  Open - 50 - 100000
  Open - 100 - 100000
  File Status - 100 - 100000
  File Status - 50 - 100000
SVT-AV1
Apache Hadoop:
  Create - 100 - 100000
  Create - 50 - 100000
oneDNN:
  IP Shapes 3D - bf16bf16bf16 - CPU
  IP Shapes 3D - f32 - CPU
  IP Shapes 3D - u8s8f32 - CPU
SVT-AV1
oneDNN:
  Convolution Batch Shapes Auto - f32 - CPU
  Convolution Batch Shapes Auto - u8s8f32 - CPU
  Convolution Batch Shapes Auto - bf16bf16bf16 - CPU
SVT-AV1:
  Preset 8 - Bosphorus 1080p
  Preset 13 - Bosphorus 4K
  Preset 12 - Bosphorus 4K
oneDNN:
  Deconvolution Batch shapes_3d - bf16bf16bf16 - CPU
  Deconvolution Batch shapes_3d - u8s8f32 - CPU
  Deconvolution Batch shapes_3d - f32 - CPU
easyWave
SVT-AV1:
  Preset 12 - Bosphorus 1080p
  Preset 13 - Bosphorus 1080p