test

lxc testing on Debian GNU/Linux 12 via the Phoronix Test Suite.

HTML result view exported from: https://openbenchmarking.org/result/2411144-NE-TEST4170061&gru&sro.

testProcessorMotherboardMemoryDiskGraphicsOSKernelCompilerFile-SystemScreen ResolutionSystem Layerr12 x Intel Xeon E5-2680 v42 x Intel Xeon E5-2680 v4 - mgag200drmfb - Dell2 x Intel Xeon E5-2680 v4 @ 3.30GHz (28 Cores / 52 Threads)Dell PowerEdge R630 02C2CP (2.19.0 BIOS)98GB1000GB TOSHIBA MQ01ABD1mgag200drmfbDebian GNU/Linux 126.8.12-3-pve (x86_64)GCC 12.2.0ext41600x1200lxc2 x Intel Xeon E5-2680 v4 @ 3.30GHz (28 Cores / 56 Threads)OpenBenchmarking.orgKernel Details- Transparent Huge Pages: madviseCompiler Details- r1, 2 x Intel Xeon E5-2680 v4: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-12-bTRWOB/gcc-12-12.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-12-bTRWOB/gcc-12-12.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- Scaling Governor: intel_cpufreq performance - CPU Microcode: 0xb000040Security Details- gather_data_sampling: Not affected + itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Mitigation of PTE Inversion; VMX: conditional cache flushes SMT vulnerable + mds: Mitigation of Clear buffers; SMT vulnerable + meltdown: Mitigation of PTI + mmio_stale_data: Mitigation of Clear buffers; SMT vulnerable + reg_file_data_sampling: Not affected + retbleed: Not affected + spec_rstack_overflow: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines; IBPB: conditional; IBRS_FW; STIBP: conditional; RSB filling; PBRSB-eIBRS: Not affected; BHI: Not affected + srbds: Not affected + tsx_async_abort: Mitigation of Clear buffers; SMT vulnerable Java Details- 2 x Intel Xeon E5-2680 v4: OpenJDK Runtime Environment (build 17.0.13+11-Debian-2deb12u1)Python Details- 2 x Intel Xeon E5-2680 v4: Python 3.11.2

testpytorch: CPU - 1 - ResNet-50pytorch: CPU - 1 - ResNet-152pytorch: CPU - 16 - ResNet-50pytorch: CPU - 32 - ResNet-50pytorch: CPU - 64 - ResNet-50pytorch: CPU - 16 - ResNet-152pytorch: CPU - 256 - ResNet-50pytorch: CPU - 32 - ResNet-152pytorch: CPU - 512 - ResNet-50pytorch: CPU - 64 - ResNet-152pytorch: CPU - 256 - ResNet-152pytorch: CPU - 512 - ResNet-152pytorch: CPU - 1 - Efficientnet_v2_lpytorch: CPU - 16 - Efficientnet_v2_lpytorch: CPU - 32 - Efficientnet_v2_lpytorch: CPU - 64 - Efficientnet_v2_lpytorch: CPU - 256 - Efficientnet_v2_lpytorch: CPU - 512 - Efficientnet_v2_lminibude: OpenMP - BM1minibude: OpenMP - BM2amg: openvino: Face Detection FP16 - CPUopenvino: Person Detection FP16 - CPUopenvino: Person Detection FP32 - CPUopenvino: Vehicle Detection FP16 - CPUopenvino: Face Detection FP16-INT8 - CPUopenvino: Face Detection Retail FP16 - CPUopenvino: Road Segmentation ADAS FP16 - CPUopenvino: Vehicle Detection FP16-INT8 - CPUopenvino: Weld Porosity Detection FP16 - CPUopenvino: Face Detection Retail FP16-INT8 - CPUopenvino: Road Segmentation ADAS FP16-INT8 - CPUopenvino: Machine Translation EN To DE FP16 - CPUopenvino: Weld Porosity Detection FP16-INT8 - CPUopenvino: Person Vehicle Bike Detection FP16 - CPUopenvino: Noise Suppression Poconet-Like FP16 - CPUopenvino: Handwritten English Recognition FP16 - CPUopenvino: Person Re-Identification Retail FP16 - CPUopenvino: Age Gender Recognition Retail 0013 FP16 - CPUopenvino: Handwritten English Recognition FP16-INT8 - CPUopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUaom-av1: Speed 0 Two-Pass - Bosphorus 4Kaom-av1: Speed 4 Two-Pass - Bosphorus 4Kaom-av1: Speed 6 Realtime - Bosphorus 4Kaom-av1: Speed 6 Two-Pass - Bosphorus 4Kaom-av1: Speed 8 Realtime - Bosphorus 4Kaom-av1: Speed 9 Realtime - Bosphorus 4Kaom-av1: Speed 10 Realtime - Bosphorus 4Kaom-av1: Speed 11 Realtime - Bosphorus 4Kaom-av1: Speed 0 Two-Pass - Bosphorus 1080paom-av1: Speed 4 Two-Pass - Bosphorus 1080paom-av1: Speed 6 Realtime - Bosphorus 1080paom-av1: Speed 6 Two-Pass - Bosphorus 1080paom-av1: Speed 8 Realtime - Bosphorus 1080paom-av1: Speed 9 Realtime - Bosphorus 1080paom-av1: Speed 10 Realtime - Bosphorus 1080paom-av1: Speed 11 Realtime - Bosphorus 1080pminibude: OpenMP - BM1minibude: OpenMP - BM2xmrig: KawPow - 1Mxmrig: Monero - 1Mxmrig: Wownero - 1Mxmrig: GhostRider - 1Mxmrig: CryptoNight-Heavy - 1Mxmrig: CryptoNight-Femto UPX2 - 1Mtensorflow: CPU - 1 - VGG-16tensorflow: GPU - 1 - VGG-16tensorflow: CPU - 1 - AlexNettensorflow: CPU - 16 - VGG-16tensorflow: CPU - 32 - VGG-16tensorflow: CPU - 64 - VGG-16tensorflow: GPU - 1 - AlexNettensorflow: GPU - 16 - VGG-16tensorflow: GPU - 32 - VGG-16tensorflow: GPU - 64 - VGG-16tensorflow: CPU - 16 - AlexNettensorflow: CPU - 256 - VGG-16tensorflow: CPU - 32 - AlexNettensorflow: CPU - 512 - VGG-16tensorflow: CPU - 64 - AlexNettensorflow: GPU - 16 - AlexNetdeepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Streamdeepsparse: NLP Document Classification, oBERT base uncased on IMDB - Synchronous Single-Streamdeepsparse: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Synchronous Single-Streamdeepsparse: ResNet-50, Baseline - Asynchronous Multi-Streamdeepsparse: ResNet-50, Baseline - Synchronous Single-Streamdeepsparse: ResNet-50, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: ResNet-50, Sparse INT8 - Synchronous Single-Streamdeepsparse: Llama2 Chat 7b Quantized - Asynchronous Multi-Streamdeepsparse: Llama2 Chat 7b Quantized - Synchronous Single-Streamdeepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Streamdeepsparse: CV Classification, ResNet-50 ImageNet - Synchronous Single-Streamdeepsparse: CV Detection, YOLOv5s COCO, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: CV Detection, YOLOv5s COCO, Sparse INT8 - Synchronous Single-Streamdeepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Streamdeepsparse: NLP Text Classification, DistilBERT mnli - Synchronous Single-Streamdeepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Streamdeepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Synchronous Single-Streamdeepsparse: BERT-Large, NLP Question Answering, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: BERT-Large, NLP Question Answering, Sparse INT8 - Synchronous Single-Streamdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Streamdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Synchronous Single-Streamgraphics-magick: Swirlgraphics-magick: Rotategraphics-magick: Sharpengraphics-magick: Enhancedgraphics-magick: Resizinggraphics-magick: Noise-Gaussiangraphics-magick: HWB Color Spaceaircrack-ng: cpuminer-opt: Magicpuminer-opt: x20rcpuminer-opt: scryptcpuminer-opt: Deepcoincpuminer-opt: Ringcoincpuminer-opt: Blake-2 Scpuminer-opt: Garlicoincpuminer-opt: Skeincoincpuminer-opt: Myriad-Groestlcpuminer-opt: LBC, LBRY Creditscpuminer-opt: Quad SHA-256, Pyritecpuminer-opt: Triple SHA-256, Onecoincachebench: Readcachebench: Writecachebench: Read / Modify / Writecompress-zstd: 3 - Compression Speedcompress-zstd: 3 - Decompression Speedcompress-zstd: 8 - Compression Speedcompress-zstd: 8 - Decompression Speedcompress-zstd: 12 - Compression Speedcompress-zstd: 12 - Decompression Speedcompress-zstd: 19 - Compression Speedcompress-zstd: 19 - Decompression Speedcompress-zstd: 3, Long Mode - Compression Speedcompress-zstd: 3, Long Mode - Decompression Speedcompress-zstd: 8, Long Mode - Compression Speedcompress-zstd: 8, Long Mode - Decompression Speedcompress-zstd: 19, Long Mode - Compression Speedcompress-zstd: 19, Long Mode - Decompression Speedhimeno: Poisson Pressure Solvercompress-7zip: Compression Ratingcompress-7zip: Decompression Ratingstockfish: Total Timelczero: BLASlczero: Eigenasmfish: 1024 Hash Memory, 26 Depthnamd: ATPase with 327,506 Atomsnamd: STMV with 1,066,628 Atomsmemcached: 1:1memcached: 1:5memcached: 5:1memcached: 1:10memcached: 1:100memtier-benchmark: Redis - 50 - 1:1memtier-benchmark: Redis - 50 - 1:5memtier-benchmark: Redis - 50 - 5:1memtier-benchmark: Redis - 100 - 1:1memtier-benchmark: Redis - 100 - 1:5memtier-benchmark: Redis - 100 - 5:1memtier-benchmark: Redis - 50 - 10:1memtier-benchmark: Redis - 50 - 1:10memtier-benchmark: Redis - 100 - 10:1memtier-benchmark: Redis - 100 - 1:10john-the-ripper: Blowfishredis: GET - 50redis: SET - 50redis: GET - 500redis: LPOP - 50redis: SADD - 50redis: SET - 500redis: GET - 1000redis: LPOP - 500redis: LPUSH - 50redis: SADD - 500redis: SET - 1000redis: LPOP - 1000redis: LPUSH - 500redis: SADD - 1000redis: LPUSH - 1000numpy: clomp: Static OMP Speedupquantlib: Squantlib: XXSllama-cpp: Meta-Llama-3-8B-Instruct-Q8_0.ggufllamafile: mistral-7b-instruct-v0.2.Q5_K_M - CPUllama-cpp: llama-2-7b.Q4_0.ggufllama-cpp: llama-2-13b.Q4_0.ggufllama-cpp: llama-2-70b-chat.Q5_0.ggufnpb: EP.Cnpb: LU.Capache-siege: 10apache-siege: 50apache-siege: 100apache-siege: 200apache-siege: 500apache-siege: 1000influxdb: 4 - 10000 - 2,5000,1 - 10000influxdb: 64 - 10000 - 2,5000,1 - 10000influxdb: 1024 - 10000 - 2,5000,1 - 10000ctx-clock: Context Switch Timepennant: sedovbigpennant: leblancbigtensorflow-lite: SqueezeNettensorflow-lite: Inception V4tensorflow-lite: NASNet Mobiletensorflow-lite: Mobilenet Floattensorflow-lite: Mobilenet Quanttensorflow-lite: Inception ResNet V2pybench: Total For Average Test Timesrenaissance: Savina Reactors.IOonednn: IP Shapes 1D - CPUonednn: IP Shapes 3D - CPUonednn: Convolution Batch Shapes Auto - CPUonednn: Deconvolution Batch shapes_1d - CPUonednn: Deconvolution Batch shapes_3d - CPUonednn: Recurrent Neural Network Training - CPUonednn: Recurrent Neural Network Inference - CPUmnn: nasnetmnn: mobilenetV3mnn: squeezenetv1.1mnn: resnet-v2-50mnn: SqueezeNetV1.0mnn: MobileNetV2_224mnn: mobilenet-v1-1.0mnn: inception-v3ncnn: CPU - mobilenetncnn: CPU-v2-v2 - mobilenet-v2ncnn: CPU-v3-v3 - mobilenet-v3ncnn: CPU - shufflenet-v2ncnn: CPU - mnasnetncnn: CPU - efficientnet-b0ncnn: CPU - blazefacencnn: CPU - googlenetncnn: CPU - vgg16ncnn: CPU - resnet18ncnn: CPU - alexnetncnn: CPU - resnet50ncnn: CPUv2-yolov3v2-yolov3 - mobilenetv2-yolov3ncnn: CPU - yolov4-tinyncnn: CPU - squeezenet_ssdncnn: CPU - regnety_400mncnn: CPU - vision_transformerncnn: CPU - FastestDetncnn: Vulkan GPU - mobilenetncnn: Vulkan GPU-v2-v2 - mobilenet-v2ncnn: Vulkan GPU-v3-v3 - mobilenet-v3ncnn: Vulkan GPU - shufflenet-v2ncnn: Vulkan GPU - mnasnetncnn: Vulkan GPU - efficientnet-b0ncnn: Vulkan GPU - blazefacencnn: Vulkan GPU - googlenetncnn: Vulkan GPU - vgg16ncnn: Vulkan GPU - resnet18ncnn: Vulkan GPU - alexnetncnn: Vulkan GPU - resnet50ncnn: Vulkan GPUv2-yolov3v2-yolov3 - mobilenetv2-yolov3ncnn: Vulkan GPU - yolov4-tinyncnn: Vulkan GPU - squeezenet_ssdncnn: Vulkan GPU - regnety_400mncnn: Vulkan GPU - vision_transformerncnn: Vulkan GPU - FastestDettnn: CPU - DenseNettnn: CPU - MobileNet v2tnn: CPU - SqueezeNet v2tnn: CPU - SqueezeNet v1.1openvino: Face Detection FP16 - CPUopenvino: Person Detection FP16 - CPUopenvino: Person Detection FP32 - CPUopenvino: Vehicle Detection FP16 - CPUopenvino: Face Detection FP16-INT8 - CPUopenvino: Face Detection Retail FP16 - CPUopenvino: Road Segmentation ADAS FP16 - CPUopenvino: Vehicle Detection FP16-INT8 - CPUopenvino: Weld Porosity Detection FP16 - CPUopenvino: Face Detection Retail FP16-INT8 - CPUopenvino: Road Segmentation ADAS FP16-INT8 - CPUopenvino: Machine Translation EN To DE FP16 - CPUopenvino: Weld Porosity Detection FP16-INT8 - CPUopenvino: Person Vehicle Bike Detection FP16 - CPUopenvino: Noise Suppression Poconet-Like FP16 - CPUopenvino: Handwritten English Recognition FP16 - CPUopenvino: Person Re-Identification Retail FP16 - CPUopenvino: Age Gender Recognition Retail 0013 FP16 - CPUopenvino: Handwritten English Recognition FP16-INT8 - CPUopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUopencv: DNN - Deep Neural Networkdeepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Streamdeepsparse: NLP Document Classification, oBERT base uncased on IMDB - Synchronous Single-Streamdeepsparse: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Synchronous Single-Streamdeepsparse: ResNet-50, Baseline - Asynchronous Multi-Streamdeepsparse: ResNet-50, Baseline - Synchronous Single-Streamdeepsparse: ResNet-50, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: ResNet-50, Sparse INT8 - Synchronous Single-Streamdeepsparse: Llama2 Chat 7b Quantized - Asynchronous Multi-Streamdeepsparse: Llama2 Chat 7b Quantized - Synchronous Single-Streamdeepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Streamdeepsparse: CV Classification, ResNet-50 ImageNet - Synchronous Single-Streamdeepsparse: CV Detection, YOLOv5s COCO, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: CV Detection, YOLOv5s COCO, Sparse INT8 - Synchronous Single-Streamdeepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Streamdeepsparse: NLP Text Classification, DistilBERT mnli - Synchronous Single-Streamdeepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Streamdeepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Synchronous Single-Streamdeepsparse: BERT-Large, NLP Question Answering, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: BERT-Large, NLP Question Answering, Sparse INT8 - Synchronous Single-Streamdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Streamdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Synchronous Single-Streamdacapobench: Jythondacapobench: Tradebeansglibc-bench: cosglibc-bench: expglibc-bench: ffsglibc-bench: powglibc-bench: singlibc-bench: log2glibc-bench: modfglibc-bench: sinhglibc-bench: sqrtglibc-bench: tanhglibc-bench: asinhglibc-bench: atanhglibc-bench: ffsllglibc-bench: sincosglibc-bench: pthread_oncerodinia: OpenMP LavaMDbuild-gcc: Time To Compilebuild-php: Time To Compilec-ray: Total Time - 4K, 16 Rays Per Pixelpovray: Trace Timerust-mandel: Time To Complete Serial/Parallel Mandelbrotcython-bench: N-Queenshackbench: 32 - Processm-queens: Time To Solverodinia: OpenMP HotSpot3Drodinia: OpenMP Leukocyterodinia: OpenMP CFD Solverrodinia: OpenMP Streamclusterincompact3d: X3D-benchmarking input.i3dincompact3d: input.i3d 129 Cells Per Directionincompact3d: input.i3d 193 Cells Per Directionopenradioss: Bumper Beamopenradioss: Cell Phone Drop Testopenradioss: Bird Strike on Windshieldblender: BMW27 - CPU-Onlyblender: Junkshop - CPU-Onlyblender: Classroom - CPU-Onlyblender: Fishy Cat - CPU-Onlyblender: Barbershop - CPU-Onlyblender: Pabellon Barcelona - CPU-Onlybuild-wasmer: Time To Compilebuild-nodejs: Time To Compileavifenc: 0avifenc: 2avifenc: 6avifenc: 6, Losslessavifenc: 10, Losslessbuild-llvm: Ninjabuild-llvm: Unix Makefilesdeepspeech: CPUrbenchmark: rnnoise: 26 Minute Long Talking Samplenumenta-nab: KNN CADnumenta-nab: Relative Entropynumenta-nab: Windowed Gaussiannumenta-nab: Earthgecko Skylinenumenta-nab: Bayesian Changepointnumenta-nab: Contextual Anomaly Detector OSEwhisper-cpp: ggml-base.en - 2016 State of the Unionwhisper-cpp: ggml-small.en - 2016 State of the Unionwhisper-cpp: ggml-medium.en - 2016 State of the Unionxnnpack: FP32MobileNetV1xnnpack: FP32MobileNetV2xnnpack: FP32MobileNetV3Largexnnpack: FP32MobileNetV3Smallxnnpack: FP16MobileNetV1xnnpack: FP16MobileNetV2xnnpack: FP16MobileNetV3Largexnnpack: FP16MobileNetV3Smallxnnpack: QS8MobileNetV2r12 x Intel Xeon E5-2680 v42 x Intel Xeon E5-2680 v4 - mgag200drmfb - Dell8361.62853936891.32970865744.17658927.4910.3222.1322.2622.008.3021.988.3422.138.408.518.465.644.354.294.324.454.2119.50620.6237026472335.2949.3350.09343.836.581379.83130.77435.20484.391299.65202.2655.26691.42422.96567.32168.09699.9915819.82214.8821265.920.215.3631.2612.0029.2931.8332.8532.910.6511.7165.1231.7066.0469.7871.4871.63487.662515.5788243.68225.412688.71582.98230.48252.72.050.797.495.996.346.493.760.850.900.9271.306.7588.376.76102.6910.8317.21868.2109345.5285100.2458194.665578.43071022.0556420.96621.28273.1765195.057378.158095.592941.5610150.082357.935115.85407.8773170.649444.911417.28638.18181548255811397513172471.815474.915936.61208.916558.282590.171242402995.14297808846.551088345963649801436.4682.6521.7668.3133.6626.511.3578.7585.4730.5495.7686.66.19592.43324.238874137422122760432610946554623665731.092680.328701268136.642992050.82809259.923191700.603195806.541524635.991609507.711361024.381365973.231480526.551303976.381331731.411517459.591277786.021481597.00306612246300.921574262.851769970.882203333.821719230.851403511.361783987.961955391.541390421.081542038.121407706.521785408.101251818.181571264.331236129.42270.3725.610.436010.94625.175.807.486.410.902736.2246814.6123642.0722369.7021617.0920413.9820425.8220260.73281287.4938427.0966559.2106338.2196021.824073967.0634692.441794.02491.033904.2042979.1116111385.63.3635512.044713.995413.16505.217061845.751018.6722.0343.5045.82526.9607.9764.9473.77238.38019.567.937.498.797.0611.223.7618.3641.4611.589.4823.7619.5632.2219.8533.5282.4010.0919.607.677.688.667.0711.313.9018.7542.4511.479.8124.0619.6032.0720.4633.7082.079.563758.318386.78493.729351.5741675.23182.19179.4226.151352.536.5068.7620.6618.556.9144.46162.5940.4621.2315.7453.4912.841.76130.161.3050567806.6228121.772440.47529.960271.854812.732013.67952.36859614.4724314.775871.708412.7763146.339324.042593.158217.2456880.9208126.921281.970722.2450805.3711122.206381441915765.509115.27824.3877034.455555.719813.15776.5817423.93995.8524632.211529.418031.98244.3861239.18325.12272133.4801491.16478.3830.61823.76245.51932.48757.43436.857141.43266.18711.15613.7471163.9602310.649378146.6904983143.2475.12230.4872.2599.22215.31105.76736.44233.4776.4092052.798150.33279.4156.73111.9278.388497.457604.370193.776030.227918.188169.27821.6178.543110.01045.69543.616202.23348543.887711539.31338229937254933367426633685511934773293OpenBenchmarking.org

PyTorch

Device: CPU - Batch Size: 1 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 1 - Model: ResNet-502 x Intel Xeon E5-2680 v4612182430SE +/- 0.28, N = 527.49MIN: 16.84 / MAX: 29.04

PyTorch

Device: CPU - Batch Size: 1 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 1 - Model: ResNet-1522 x Intel Xeon E5-2680 v43691215SE +/- 0.07, N = 1210.32MIN: 6.9 / MAX: 11.18

PyTorch

Device: CPU - Batch Size: 16 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 16 - Model: ResNet-502 x Intel Xeon E5-2680 v4510152025SE +/- 0.07, N = 322.13MIN: 19.61 / MAX: 22.61

PyTorch

Device: CPU - Batch Size: 32 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 32 - Model: ResNet-502 x Intel Xeon E5-2680 v4510152025SE +/- 0.28, N = 422.26MIN: 18.7 / MAX: 23.31

PyTorch

Device: CPU - Batch Size: 64 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 64 - Model: ResNet-502 x Intel Xeon E5-2680 v4510152025SE +/- 0.23, N = 522.00MIN: 15.47 / MAX: 22.89

PyTorch

Device: CPU - Batch Size: 16 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 16 - Model: ResNet-1522 x Intel Xeon E5-2680 v4246810SE +/- 0.06, N = 38.30MIN: 6.51 / MAX: 8.57

PyTorch

Device: CPU - Batch Size: 256 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 256 - Model: ResNet-502 x Intel Xeon E5-2680 v4510152025SE +/- 0.23, N = 1221.98MIN: 15.73 / MAX: 23.11

PyTorch

Device: CPU - Batch Size: 32 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 32 - Model: ResNet-1522 x Intel Xeon E5-2680 v4246810SE +/- 0.01, N = 38.34MIN: 6.58 / MAX: 8.48

PyTorch

Device: CPU - Batch Size: 512 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 512 - Model: ResNet-502 x Intel Xeon E5-2680 v4510152025SE +/- 0.21, N = 322.13MIN: 19.45 / MAX: 22.71

PyTorch

Device: CPU - Batch Size: 64 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 64 - Model: ResNet-1522 x Intel Xeon E5-2680 v4246810SE +/- 0.08, N = 98.40MIN: 6.58 / MAX: 8.89

PyTorch

Device: CPU - Batch Size: 256 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 256 - Model: ResNet-1522 x Intel Xeon E5-2680 v4246810SE +/- 0.12, N = 38.51MIN: 6.24 / MAX: 8.82

PyTorch

Device: CPU - Batch Size: 512 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 512 - Model: ResNet-1522 x Intel Xeon E5-2680 v4246810SE +/- 0.09, N = 98.46MIN: 1.24 / MAX: 8.92

PyTorch

Device: CPU - Batch Size: 1 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 1 - Model: Efficientnet_v2_l2 x Intel Xeon E5-2680 v41.2692.5383.8075.0766.345SE +/- 0.04, N = 35.64MIN: 4.26 / MAX: 6.2

PyTorch

Device: CPU - Batch Size: 16 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 16 - Model: Efficientnet_v2_l2 x Intel Xeon E5-2680 v40.97881.95762.93643.91524.894SE +/- 0.04, N = 94.35MIN: 3.6 / MAX: 4.55

PyTorch

Device: CPU - Batch Size: 32 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 32 - Model: Efficientnet_v2_l2 x Intel Xeon E5-2680 v40.96531.93062.89593.86124.8265SE +/- 0.04, N = 94.29MIN: 3.33 / MAX: 4.53

PyTorch

Device: CPU - Batch Size: 64 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 64 - Model: Efficientnet_v2_l2 x Intel Xeon E5-2680 v40.9721.9442.9163.8884.86SE +/- 0.06, N = 94.32MIN: 3.4 / MAX: 4.6

PyTorch

Device: CPU - Batch Size: 256 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 256 - Model: Efficientnet_v2_l2 x Intel Xeon E5-2680 v41.00132.00263.00394.00525.0065SE +/- 0.04, N = 34.45MIN: 3.4 / MAX: 4.59

PyTorch

Device: CPU - Batch Size: 512 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.2.1Device: CPU - Batch Size: 512 - Model: Efficientnet_v2_l2 x Intel Xeon E5-2680 v40.94731.89462.84193.78924.7365SE +/- 0.04, N = 34.21MIN: 3.28 / MAX: 4.35

miniBUDE

Implementation: OpenMP - Input Deck: BM1

OpenBenchmarking.orgBillion Interactions/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM12 x Intel Xeon E5-2680 v4510152025SE +/- 0.09, N = 319.511. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm

miniBUDE

Implementation: OpenMP - Input Deck: BM2

OpenBenchmarking.orgBillion Interactions/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM22 x Intel Xeon E5-2680 v4510152025SE +/- 0.02, N = 320.621. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm

Algebraic Multi-Grid Benchmark

OpenBenchmarking.orgFigure Of Merit, More Is BetterAlgebraic Multi-Grid Benchmark 1.22 x Intel Xeon E5-2680 v4150M300M450M600M750MSE +/- 3355643.09, N = 37026472331. (CC) gcc options: -lparcsr_ls -lparcsr_mv -lseq_mv -lIJ_mv -lkrylov -lHYPRE_utilities -lm -fopenmp -lmpi

OpenVINO

Model: Face Detection FP16 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Face Detection FP16 - Device: CPU2 x Intel Xeon E5-2680 v41.19032.38063.57094.76125.9515SE +/- 0.01, N = 35.291. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Person Detection FP16 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Person Detection FP16 - Device: CPU2 x Intel Xeon E5-2680 v41122334455SE +/- 0.27, N = 349.331. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Person Detection FP32 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Person Detection FP32 - Device: CPU2 x Intel Xeon E5-2680 v41122334455SE +/- 0.33, N = 350.091. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Vehicle Detection FP16 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Vehicle Detection FP16 - Device: CPU2 x Intel Xeon E5-2680 v470140210280350SE +/- 4.27, N = 3343.831. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Face Detection FP16-INT8 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Face Detection FP16-INT8 - Device: CPU2 x Intel Xeon E5-2680 v4246810SE +/- 0.01, N = 36.581. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Face Detection Retail FP16 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Face Detection Retail FP16 - Device: CPU2 x Intel Xeon E5-2680 v430060090012001500SE +/- 4.45, N = 31379.831. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Road Segmentation ADAS FP16 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Road Segmentation ADAS FP16 - Device: CPU2 x Intel Xeon E5-2680 v4306090120150SE +/- 0.72, N = 3130.771. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Vehicle Detection FP16-INT8 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Vehicle Detection FP16-INT8 - Device: CPU2 x Intel Xeon E5-2680 v490180270360450SE +/- 0.12, N = 3435.201. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Weld Porosity Detection FP16 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Weld Porosity Detection FP16 - Device: CPU2 x Intel Xeon E5-2680 v4100200300400500SE +/- 0.11, N = 3484.391. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Face Detection Retail FP16-INT8 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Face Detection Retail FP16-INT8 - Device: CPU2 x Intel Xeon E5-2680 v430060090012001500SE +/- 0.44, N = 31299.651. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Road Segmentation ADAS FP16-INT8 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Road Segmentation ADAS FP16-INT8 - Device: CPU2 x Intel Xeon E5-2680 v44080120160200SE +/- 0.29, N = 3202.261. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Machine Translation EN To DE FP16 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Machine Translation EN To DE FP16 - Device: CPU2 x Intel Xeon E5-2680 v41224364860SE +/- 0.22, N = 355.261. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Weld Porosity Detection FP16-INT8 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Weld Porosity Detection FP16-INT8 - Device: CPU2 x Intel Xeon E5-2680 v4150300450600750SE +/- 0.53, N = 3691.421. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Person Vehicle Bike Detection FP16 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Person Vehicle Bike Detection FP16 - Device: CPU2 x Intel Xeon E5-2680 v490180270360450SE +/- 0.06, N = 3422.961. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Noise Suppression Poconet-Like FP16 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Noise Suppression Poconet-Like FP16 - Device: CPU2 x Intel Xeon E5-2680 v4120240360480600SE +/- 0.35, N = 3567.321. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Handwritten English Recognition FP16 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Handwritten English Recognition FP16 - Device: CPU2 x Intel Xeon E5-2680 v44080120160200SE +/- 0.13, N = 3168.091. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Person Re-Identification Retail FP16 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Person Re-Identification Retail FP16 - Device: CPU2 x Intel Xeon E5-2680 v4150300450600750SE +/- 0.71, N = 3699.991. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU2 x Intel Xeon E5-2680 v43K6K9K12K15KSE +/- 56.67, N = 315819.821. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Handwritten English Recognition FP16-INT8 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Handwritten English Recognition FP16-INT8 - Device: CPU2 x Intel Xeon E5-2680 v450100150200250SE +/- 0.61, N = 3214.881. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2024.0Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU2 x Intel Xeon E5-2680 v45K10K15K20K25KSE +/- 19.67, N = 321265.921. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

AOM AV1

Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 4K

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 4K2 x Intel Xeon E5-2680 v40.04730.09460.14190.18920.2365SE +/- 0.00, N = 30.211. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

AOM AV1

Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 4K

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 4K2 x Intel Xeon E5-2680 v41.2062.4123.6184.8246.03SE +/- 0.03, N = 35.361. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

AOM AV1

Encoder Mode: Speed 6 Realtime - Input: Bosphorus 4K

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 6 Realtime - Input: Bosphorus 4K2 x Intel Xeon E5-2680 v4714212835SE +/- 0.24, N = 1531.261. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

AOM AV1

Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 4K

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 4K2 x Intel Xeon E5-2680 v43691215SE +/- 0.04, N = 312.001. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

AOM AV1

Encoder Mode: Speed 8 Realtime - Input: Bosphorus 4K

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 8 Realtime - Input: Bosphorus 4K2 x Intel Xeon E5-2680 v4714212835SE +/- 0.19, N = 329.291. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

AOM AV1

Encoder Mode: Speed 9 Realtime - Input: Bosphorus 4K

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 9 Realtime - Input: Bosphorus 4K2 x Intel Xeon E5-2680 v4714212835SE +/- 0.31, N = 331.831. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

AOM AV1

Encoder Mode: Speed 10 Realtime - Input: Bosphorus 4K

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 10 Realtime - Input: Bosphorus 4K2 x Intel Xeon E5-2680 v4816243240SE +/- 0.34, N = 332.851. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

AOM AV1

Encoder Mode: Speed 11 Realtime - Input: Bosphorus 4K

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 11 Realtime - Input: Bosphorus 4K2 x Intel Xeon E5-2680 v4816243240SE +/- 0.13, N = 332.911. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

AOM AV1

Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 1080p

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 1080p2 x Intel Xeon E5-2680 v40.14630.29260.43890.58520.7315SE +/- 0.01, N = 30.651. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

AOM AV1

Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 1080p

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 1080p2 x Intel Xeon E5-2680 v43691215SE +/- 0.08, N = 311.711. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

AOM AV1

Encoder Mode: Speed 6 Realtime - Input: Bosphorus 1080p

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 6 Realtime - Input: Bosphorus 1080p2 x Intel Xeon E5-2680 v41530456075SE +/- 0.49, N = 1565.121. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

AOM AV1

Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 1080p

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 1080p2 x Intel Xeon E5-2680 v4714212835SE +/- 0.28, N = 1531.701. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

AOM AV1

Encoder Mode: Speed 8 Realtime - Input: Bosphorus 1080p

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 8 Realtime - Input: Bosphorus 1080p2 x Intel Xeon E5-2680 v41530456075SE +/- 0.85, N = 1566.041. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

AOM AV1

Encoder Mode: Speed 9 Realtime - Input: Bosphorus 1080p

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 9 Realtime - Input: Bosphorus 1080p2 x Intel Xeon E5-2680 v41632486480SE +/- 1.11, N = 1569.781. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

AOM AV1

Encoder Mode: Speed 10 Realtime - Input: Bosphorus 1080p

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 10 Realtime - Input: Bosphorus 1080p2 x Intel Xeon E5-2680 v41632486480SE +/- 0.76, N = 1571.481. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

AOM AV1

Encoder Mode: Speed 11 Realtime - Input: Bosphorus 1080p

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.9Encoder Mode: Speed 11 Realtime - Input: Bosphorus 1080p2 x Intel Xeon E5-2680 v41632486480SE +/- 0.80, N = 1571.631. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

miniBUDE

Implementation: OpenMP - Input Deck: BM1

OpenBenchmarking.orgGFInst/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM12 x Intel Xeon E5-2680 v4110220330440550SE +/- 2.27, N = 3487.661. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm

miniBUDE

Implementation: OpenMP - Input Deck: BM2

OpenBenchmarking.orgGFInst/s, More Is BetterminiBUDE 20210901Implementation: OpenMP - Input Deck: BM22 x Intel Xeon E5-2680 v4110220330440550SE +/- 0.62, N = 3515.581. (CC) gcc options: -std=c99 -Ofast -ffast-math -fopenmp -march=native -lm

Xmrig

Variant: KawPow - Hash Count: 1M

OpenBenchmarking.orgH/s, More Is BetterXmrig 6.21Variant: KawPow - Hash Count: 1M2 x Intel Xeon E5-2680 v42K4K6K8K10KSE +/- 18.01, N = 38243.61. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc

Xmrig

Variant: Monero - Hash Count: 1M

OpenBenchmarking.orgH/s, More Is BetterXmrig 6.21Variant: Monero - Hash Count: 1M2 x Intel Xeon E5-2680 v42K4K6K8K10KSE +/- 33.81, N = 38225.41. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc

Xmrig

Variant: Wownero - Hash Count: 1M

OpenBenchmarking.orgH/s, More Is BetterXmrig 6.21Variant: Wownero - Hash Count: 1M2 x Intel Xeon E5-2680 v43K6K9K12K15KSE +/- 82.76, N = 312688.71. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc

Xmrig

Variant: GhostRider - Hash Count: 1M

OpenBenchmarking.orgH/s, More Is BetterXmrig 6.21Variant: GhostRider - Hash Count: 1M2 x Intel Xeon E5-2680 v430060090012001500SE +/- 0.69, N = 31582.91. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc

Xmrig

Variant: CryptoNight-Heavy - Hash Count: 1M

OpenBenchmarking.orgH/s, More Is BetterXmrig 6.21Variant: CryptoNight-Heavy - Hash Count: 1M2 x Intel Xeon E5-2680 v42K4K6K8K10KSE +/- 29.52, N = 38230.41. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc

Xmrig

Variant: CryptoNight-Femto UPX2 - Hash Count: 1M

OpenBenchmarking.orgH/s, More Is BetterXmrig 6.21Variant: CryptoNight-Femto UPX2 - Hash Count: 1M2 x Intel Xeon E5-2680 v42K4K6K8K10KSE +/- 30.75, N = 38252.71. (CXX) g++ options: -fexceptions -fno-rtti -maes -O3 -Ofast -static-libgcc -static-libstdc++ -rdynamic -lssl -lcrypto -luv -lpthread -lrt -ldl -lhwloc

TensorFlow

Device: CPU - Batch Size: 1 - Model: VGG-16

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: CPU - Batch Size: 1 - Model: VGG-162 x Intel Xeon E5-2680 v40.46130.92261.38391.84522.3065SE +/- 0.02, N = 32.05

TensorFlow

Device: GPU - Batch Size: 1 - Model: VGG-16

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: GPU - Batch Size: 1 - Model: VGG-162 x Intel Xeon E5-2680 v40.17780.35560.53340.71120.889SE +/- 0.00, N = 30.79

TensorFlow

Device: CPU - Batch Size: 1 - Model: AlexNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: CPU - Batch Size: 1 - Model: AlexNet2 x Intel Xeon E5-2680 v4246810SE +/- 0.06, N = 37.49

TensorFlow

Device: CPU - Batch Size: 16 - Model: VGG-16

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: CPU - Batch Size: 16 - Model: VGG-162 x Intel Xeon E5-2680 v41.34782.69564.04345.39126.739SE +/- 0.02, N = 35.99

TensorFlow

Device: CPU - Batch Size: 32 - Model: VGG-16

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: CPU - Batch Size: 32 - Model: VGG-162 x Intel Xeon E5-2680 v4246810SE +/- 0.01, N = 36.34

TensorFlow

Device: CPU - Batch Size: 64 - Model: VGG-16

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: CPU - Batch Size: 64 - Model: VGG-162 x Intel Xeon E5-2680 v4246810SE +/- 0.01, N = 36.49

TensorFlow

Device: GPU - Batch Size: 1 - Model: AlexNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: GPU - Batch Size: 1 - Model: AlexNet2 x Intel Xeon E5-2680 v40.8461.6922.5383.3844.23SE +/- 0.02, N = 33.76

TensorFlow

Device: GPU - Batch Size: 16 - Model: VGG-16

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: GPU - Batch Size: 16 - Model: VGG-162 x Intel Xeon E5-2680 v40.19130.38260.57390.76520.9565SE +/- 0.00, N = 30.85

TensorFlow

Device: GPU - Batch Size: 32 - Model: VGG-16

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: GPU - Batch Size: 32 - Model: VGG-162 x Intel Xeon E5-2680 v40.20250.4050.60750.811.0125SE +/- 0.01, N = 30.90

TensorFlow

Device: GPU - Batch Size: 64 - Model: VGG-16

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: GPU - Batch Size: 64 - Model: VGG-162 x Intel Xeon E5-2680 v40.2070.4140.6210.8281.0350.92

TensorFlow

Device: CPU - Batch Size: 16 - Model: AlexNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: CPU - Batch Size: 16 - Model: AlexNet2 x Intel Xeon E5-2680 v41632486480SE +/- 0.21, N = 371.30

TensorFlow

Device: CPU - Batch Size: 256 - Model: VGG-16

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: CPU - Batch Size: 256 - Model: VGG-162 x Intel Xeon E5-2680 v4246810SE +/- 0.01, N = 36.75

TensorFlow

Device: CPU - Batch Size: 32 - Model: AlexNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: CPU - Batch Size: 32 - Model: AlexNet2 x Intel Xeon E5-2680 v420406080100SE +/- 0.90, N = 388.37

TensorFlow

Device: CPU - Batch Size: 512 - Model: VGG-16

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: CPU - Batch Size: 512 - Model: VGG-162 x Intel Xeon E5-2680 v4246810SE +/- 0.03, N = 36.76

TensorFlow

Device: CPU - Batch Size: 64 - Model: AlexNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: CPU - Batch Size: 64 - Model: AlexNet2 x Intel Xeon E5-2680 v420406080100SE +/- 0.40, N = 3102.69

TensorFlow

Device: GPU - Batch Size: 16 - Model: AlexNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.16.1Device: GPU - Batch Size: 16 - Model: AlexNet2 x Intel Xeon E5-2680 v43691215SE +/- 0.06, N = 310.83

Neural Magic DeepSparse

Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v448121620SE +/- 0.04, N = 317.22

Neural Magic DeepSparse

Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v4246810SE +/- 0.0110, N = 38.2109

Neural Magic DeepSparse

Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v480160240320400SE +/- 0.63, N = 3345.53

Neural Magic DeepSparse

Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v420406080100SE +/- 0.28, N = 3100.25

Neural Magic DeepSparse

Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v44080120160200SE +/- 0.67, N = 3194.67

Neural Magic DeepSparse

Model: ResNet-50, Baseline - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: ResNet-50, Baseline - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v420406080100SE +/- 0.05, N = 378.43

Neural Magic DeepSparse

Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v42004006008001000SE +/- 2.18, N = 31022.06

Neural Magic DeepSparse

Model: ResNet-50, Sparse INT8 - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: ResNet-50, Sparse INT8 - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v490180270360450SE +/- 1.03, N = 3420.97

Neural Magic DeepSparse

Model: Llama2 Chat 7b Quantized - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: Llama2 Chat 7b Quantized - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v40.28860.57720.86581.15441.443SE +/- 0.0121, N = 31.2827

Neural Magic DeepSparse

Model: Llama2 Chat 7b Quantized - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: Llama2 Chat 7b Quantized - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v40.71471.42942.14412.85883.5735SE +/- 0.0072, N = 33.1765

Neural Magic DeepSparse

Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v44080120160200SE +/- 0.90, N = 3195.06

Neural Magic DeepSparse

Model: CV Classification, ResNet-50 ImageNet - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: CV Classification, ResNet-50 ImageNet - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v420406080100SE +/- 0.13, N = 378.16

Neural Magic DeepSparse

Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v420406080100SE +/- 0.23, N = 395.59

Neural Magic DeepSparse

Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v4918273645SE +/- 0.11, N = 341.56

Neural Magic DeepSparse

Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v4306090120150SE +/- 0.36, N = 3150.08

Neural Magic DeepSparse

Model: NLP Text Classification, DistilBERT mnli - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: NLP Text Classification, DistilBERT mnli - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v41326395265SE +/- 0.25, N = 357.94

Neural Magic DeepSparse

Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v448121620SE +/- 0.02, N = 315.85

Neural Magic DeepSparse

Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v4246810SE +/- 0.0094, N = 37.8773

Neural Magic DeepSparse

Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v44080120160200SE +/- 0.13, N = 3170.65

Neural Magic DeepSparse

Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v41020304050SE +/- 0.13, N = 344.91

Neural Magic DeepSparse

Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v448121620SE +/- 0.04, N = 317.29

Neural Magic DeepSparse

Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.7Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v4246810SE +/- 0.0178, N = 38.1818

GraphicsMagick

Operation: Swirl

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Swirl2 x Intel Xeon E5-2680 v4306090120150SE +/- 3.41, N = 121541. (CC) gcc options: -fopenmp -O2 -ltiff -ljbig -lwebpmux -lwebp -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -lxml2 -lzstd -llzma -lz -lm -lpthread -lgomp

GraphicsMagick

Operation: Rotate

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Rotate2 x Intel Xeon E5-2680 v420406080100SE +/- 0.00, N = 3821. (CC) gcc options: -fopenmp -O2 -ltiff -ljbig -lwebpmux -lwebp -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -lxml2 -lzstd -llzma -lz -lm -lpthread -lgomp

GraphicsMagick

Operation: Sharpen

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Sharpen2 x Intel Xeon E5-2680 v41224364860SE +/- 0.33, N = 3551. (CC) gcc options: -fopenmp -O2 -ltiff -ljbig -lwebpmux -lwebp -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -lxml2 -lzstd -llzma -lz -lm -lpthread -lgomp

GraphicsMagick

Operation: Enhanced

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Enhanced2 x Intel Xeon E5-2680 v420406080100SE +/- 0.00, N = 3811. (CC) gcc options: -fopenmp -O2 -ltiff -ljbig -lwebpmux -lwebp -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -lxml2 -lzstd -llzma -lz -lm -lpthread -lgomp

GraphicsMagick

Operation: Resizing

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Resizing2 x Intel Xeon E5-2680 v4306090120150SE +/- 2.05, N = 151391. (CC) gcc options: -fopenmp -O2 -ltiff -ljbig -lwebpmux -lwebp -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -lxml2 -lzstd -llzma -lz -lm -lpthread -lgomp

GraphicsMagick

Operation: Noise-Gaussian

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Noise-Gaussian2 x Intel Xeon E5-2680 v420406080100SE +/- 0.33, N = 3751. (CC) gcc options: -fopenmp -O2 -ltiff -ljbig -lwebpmux -lwebp -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -lxml2 -lzstd -llzma -lz -lm -lpthread -lgomp

GraphicsMagick

Operation: HWB Color Space

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: HWB Color Space2 x Intel Xeon E5-2680 v4306090120150SE +/- 0.33, N = 31311. (CC) gcc options: -fopenmp -O2 -ltiff -ljbig -lwebpmux -lwebp -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -lxml2 -lzstd -llzma -lz -lm -lpthread -lgomp

Aircrack-ng

OpenBenchmarking.orgk/s, More Is BetterAircrack-ng 1.72 x Intel Xeon E5-2680 v416K32K48K64K80KSE +/- 97.52, N = 372471.821. (CXX) g++ options: -std=gnu++17 -O3 -fvisibility=hidden -fcommon -rdynamic -lnl-3 -lnl-genl-3 -lpcre -lpthread -lz -lssl -lcrypto -lhwloc -ldl -lm -pthread

Cpuminer-Opt

Algorithm: Magi

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 24.3Algorithm: Magi2 x Intel Xeon E5-2680 v4100200300400500SE +/- 0.82, N = 3474.911. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lgmp

Cpuminer-Opt

Algorithm: x20r

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 24.3Algorithm: x20r2 x Intel Xeon E5-2680 v413002600390052006500SE +/- 2.22, N = 35936.611. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lgmp

Cpuminer-Opt

Algorithm: scrypt

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 24.3Algorithm: scrypt2 x Intel Xeon E5-2680 v450100150200250SE +/- 0.14, N = 3208.911. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lgmp

Cpuminer-Opt

Algorithm: Deepcoin

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 24.3Algorithm: Deepcoin2 x Intel Xeon E5-2680 v414002800420056007000SE +/- 2.36, N = 36558.281. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lgmp

Cpuminer-Opt

Algorithm: Ringcoin

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 24.3Algorithm: Ringcoin2 x Intel Xeon E5-2680 v46001200180024003000SE +/- 9.31, N = 32590.171. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lgmp

Cpuminer-Opt

Algorithm: Blake-2 S

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 24.3Algorithm: Blake-2 S2 x Intel Xeon E5-2680 v430K60K90K120K150KSE +/- 210.08, N = 31242401. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lgmp

Cpuminer-Opt

Algorithm: Garlicoin

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 24.3Algorithm: Garlicoin2 x Intel Xeon E5-2680 v46001200180024003000SE +/- 17.79, N = 32995.141. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lgmp

Cpuminer-Opt

Algorithm: Skeincoin

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 24.3Algorithm: Skeincoin2 x Intel Xeon E5-2680 v46K12K18K24K30KSE +/- 215.02, N = 3297801. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lgmp

Cpuminer-Opt

Algorithm: Myriad-Groestl

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 24.3Algorithm: Myriad-Groestl2 x Intel Xeon E5-2680 v42K4K6K8K10KSE +/- 7.21, N = 38846.551. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lgmp

Cpuminer-Opt

Algorithm: LBC, LBRY Credits

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 24.3Algorithm: LBC, LBRY Credits2 x Intel Xeon E5-2680 v42K4K6K8K10KSE +/- 3.33, N = 3108831. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lgmp

Cpuminer-Opt

Algorithm: Quad SHA-256, Pyrite

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 24.3Algorithm: Quad SHA-256, Pyrite2 x Intel Xeon E5-2680 v410K20K30K40K50KSE +/- 459.94, N = 3459631. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lgmp

Cpuminer-Opt

Algorithm: Triple SHA-256, Onecoin

OpenBenchmarking.orgkH/s, More Is BetterCpuminer-Opt 24.3Algorithm: Triple SHA-256, Onecoin2 x Intel Xeon E5-2680 v414K28K42K56K70KSE +/- 5.77, N = 3649801. (CXX) g++ options: -O2 -lcurl -lz -lpthread -lgmp

CacheBench

Test: Read

OpenBenchmarking.orgMB/s, More Is BetterCacheBenchTest: Readr12K4K6K8K10KSE +/- 0.03, N = 38361.63MIN: 8349.48 / MAX: 8369.761. (CC) gcc options: -O3 -lrt

CacheBench

Test: Write

OpenBenchmarking.orgMB/s, More Is BetterCacheBenchTest: Writer18K16K24K32K40KSE +/- 391.50, N = 536891.33MIN: 22723.54 / MAX: 48193.41. (CC) gcc options: -O3 -lrt

CacheBench

Test: Read / Modify / Write

OpenBenchmarking.orgMB/s, More Is BetterCacheBenchTest: Read / Modify / Writer114K28K42K56K70KSE +/- 147.90, N = 365744.18MIN: 46193.67 / MAX: 78582.681. (CC) gcc options: -O3 -lrt

Zstd Compression

Compression Level: 3 - Compression Speed

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3 - Compression Speed2 x Intel Xeon E5-2680 v430060090012001500SE +/- 8.07, N = 31436.41. (CC) gcc options: -O3 -pthread -lz -llzma

Zstd Compression

Compression Level: 3 - Decompression Speed

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3 - Decompression Speed2 x Intel Xeon E5-2680 v4150300450600750SE +/- 1.71, N = 3682.61. (CC) gcc options: -O3 -pthread -lz -llzma

Zstd Compression

Compression Level: 8 - Compression Speed

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8 - Compression Speed2 x Intel Xeon E5-2680 v4110220330440550SE +/- 5.93, N = 3521.71. (CC) gcc options: -O3 -pthread -lz -llzma

Zstd Compression

Compression Level: 8 - Decompression Speed

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8 - Decompression Speed2 x Intel Xeon E5-2680 v4140280420560700SE +/- 1.84, N = 3668.31. (CC) gcc options: -O3 -pthread -lz -llzma

Zstd Compression

Compression Level: 12 - Compression Speed

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 12 - Compression Speed2 x Intel Xeon E5-2680 v4306090120150SE +/- 1.42, N = 3133.61. (CC) gcc options: -O3 -pthread -lz -llzma

Zstd Compression

Compression Level: 12 - Decompression Speed

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 12 - Decompression Speed2 x Intel Xeon E5-2680 v4140280420560700SE +/- 0.86, N = 3626.51. (CC) gcc options: -O3 -pthread -lz -llzma

Zstd Compression

Compression Level: 19 - Compression Speed

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19 - Compression Speed2 x Intel Xeon E5-2680 v43691215SE +/- 0.07, N = 311.31. (CC) gcc options: -O3 -pthread -lz -llzma

Zstd Compression

Compression Level: 19 - Decompression Speed

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19 - Decompression Speed2 x Intel Xeon E5-2680 v4130260390520650SE +/- 0.74, N = 3578.71. (CC) gcc options: -O3 -pthread -lz -llzma

Zstd Compression

Compression Level: 3, Long Mode - Compression Speed

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3, Long Mode - Compression Speed2 x Intel Xeon E5-2680 v4130260390520650SE +/- 3.88, N = 3585.41. (CC) gcc options: -O3 -pthread -lz -llzma

Zstd Compression

Compression Level: 3, Long Mode - Decompression Speed

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3, Long Mode - Decompression Speed2 x Intel Xeon E5-2680 v4160320480640800SE +/- 2.39, N = 3730.51. (CC) gcc options: -O3 -pthread -lz -llzma

Zstd Compression

Compression Level: 8, Long Mode - Compression Speed

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8, Long Mode - Compression Speed2 x Intel Xeon E5-2680 v4110220330440550SE +/- 1.63, N = 3495.71. (CC) gcc options: -O3 -pthread -lz -llzma

Zstd Compression

Compression Level: 8, Long Mode - Decompression Speed

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8, Long Mode - Decompression Speed2 x Intel Xeon E5-2680 v4150300450600750SE +/- 0.21, N = 3686.61. (CC) gcc options: -O3 -pthread -lz -llzma

Zstd Compression

Compression Level: 19, Long Mode - Compression Speed

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19, Long Mode - Compression Speed2 x Intel Xeon E5-2680 v4246810SE +/- 0.08, N = 36.191. (CC) gcc options: -O3 -pthread -lz -llzma

Zstd Compression

Compression Level: 19, Long Mode - Decompression Speed

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19, Long Mode - Decompression Speed2 x Intel Xeon E5-2680 v4130260390520650SE +/- 0.40, N = 3592.41. (CC) gcc options: -O3 -pthread -lz -llzma

Himeno Benchmark

Poisson Pressure Solver

OpenBenchmarking.orgMFLOPS, More Is BetterHimeno Benchmark 3.0Poisson Pressure Solver2 x Intel Xeon E5-2680 v47001400210028003500SE +/- 6.30, N = 33324.241. (CC) gcc options: -O3 -mavx2

7-Zip Compression

Test: Compression Rating

OpenBenchmarking.orgMIPS, More Is Better7-Zip Compression 24.05Test: Compression Rating2 x Intel Xeon E5-2680 v430K60K90K120K150KSE +/- 904.38, N = 31374221. (CXX) g++ options: -lpthread -ldl -O2 -fPIC

7-Zip Compression

Test: Decompression Rating

OpenBenchmarking.orgMIPS, More Is Better7-Zip Compression 24.05Test: Decompression Rating2 x Intel Xeon E5-2680 v430K60K90K120K150KSE +/- 70.29, N = 31227601. (CXX) g++ options: -lpthread -ldl -O2 -fPIC

Stockfish

Total Time

OpenBenchmarking.orgNodes Per Second, More Is BetterStockfish 17Total Time2 x Intel Xeon E5-2680 v49M18M27M36M45MSE +/- 1759062.68, N = 6432610941. (CXX) g++ options: -lgcov -m64 -lpthread -fno-exceptions -std=c++17 -fno-peel-loops -fno-tracer -pedantic -O3 -funroll-loops -msse -msse3 -mpopcnt -mavx2 -mbmi -msse4.1 -mssse3 -msse2 -mbmi2 -flto -flto-partition=one -flto=jobserver

LeelaChessZero

Backend: BLAS

OpenBenchmarking.orgNodes Per Second, More Is BetterLeelaChessZero 0.31.1Backend: BLAS2 x Intel Xeon E5-2680 v41530456075SE +/- 1.15, N = 9651. (CXX) g++ options: -flto -pthread

LeelaChessZero

Backend: Eigen

OpenBenchmarking.orgNodes Per Second, More Is BetterLeelaChessZero 0.31.1Backend: Eigen2 x Intel Xeon E5-2680 v41224364860SE +/- 1.37, N = 9541. (CXX) g++ options: -flto -pthread

asmFish

1024 Hash Memory, 26 Depth

OpenBenchmarking.orgNodes/second, More Is BetterasmFish 2018-07-231024 Hash Memory, 26 Depth2 x Intel Xeon E5-2680 v413M26M39M52M65MSE +/- 347481.30, N = 362366573

NAMD

Input: ATPase with 327,506 Atoms

OpenBenchmarking.orgns/day, More Is BetterNAMD 3.0Input: ATPase with 327,506 Atoms2 x Intel Xeon E5-2680 v40.24590.49180.73770.98361.2295SE +/- 0.00795, N = 31.09268

NAMD

Input: STMV with 1,066,628 Atoms

OpenBenchmarking.orgns/day, More Is BetterNAMD 3.0Input: STMV with 1,066,628 Atoms2 x Intel Xeon E5-2680 v40.0740.1480.2220.2960.37SE +/- 0.00061, N = 30.32870

Memcached

Set To Get Ratio: 1:1

OpenBenchmarking.orgOps/sec, More Is BetterMemcached 1.6.19Set To Get Ratio: 1:12 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 16311.34, N = 31268136.641. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

Memcached

Set To Get Ratio: 1:5

OpenBenchmarking.orgOps/sec, More Is BetterMemcached 1.6.19Set To Get Ratio: 1:52 x Intel Xeon E5-2680 v4600K1200K1800K2400K3000KSE +/- 13235.65, N = 32992050.821. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

Memcached

Set To Get Ratio: 5:1

OpenBenchmarking.orgOps/sec, More Is BetterMemcached 1.6.19Set To Get Ratio: 5:12 x Intel Xeon E5-2680 v4200K400K600K800K1000KSE +/- 3234.70, N = 3809259.921. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

Memcached

Set To Get Ratio: 1:10

OpenBenchmarking.orgOps/sec, More Is BetterMemcached 1.6.19Set To Get Ratio: 1:102 x Intel Xeon E5-2680 v4700K1400K2100K2800K3500KSE +/- 2765.51, N = 33191700.601. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

Memcached

Set To Get Ratio: 1:100

OpenBenchmarking.orgOps/sec, More Is BetterMemcached 1.6.19Set To Get Ratio: 1:1002 x Intel Xeon E5-2680 v4700K1400K2100K2800K3500KSE +/- 26780.92, N = 33195806.541. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

Redis 7.0.12 + memtier_benchmark

Protocol: Redis - Clients: 50 - Set To Get Ratio: 1:1

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 50 - Set To Get Ratio: 1:12 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 4640.23, N = 31524635.991. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

Redis 7.0.12 + memtier_benchmark

Protocol: Redis - Clients: 50 - Set To Get Ratio: 1:5

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 50 - Set To Get Ratio: 1:52 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 3714.54, N = 31609507.711. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

Redis 7.0.12 + memtier_benchmark

Protocol: Redis - Clients: 50 - Set To Get Ratio: 5:1

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 50 - Set To Get Ratio: 5:12 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 2137.82, N = 31361024.381. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

Redis 7.0.12 + memtier_benchmark

Protocol: Redis - Clients: 100 - Set To Get Ratio: 1:1

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 100 - Set To Get Ratio: 1:12 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 4068.22, N = 31365973.231. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

Redis 7.0.12 + memtier_benchmark

Protocol: Redis - Clients: 100 - Set To Get Ratio: 1:5

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 100 - Set To Get Ratio: 1:52 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 7788.01, N = 31480526.551. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

Redis 7.0.12 + memtier_benchmark

Protocol: Redis - Clients: 100 - Set To Get Ratio: 5:1

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 100 - Set To Get Ratio: 5:12 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 3475.92, N = 31303976.381. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

Redis 7.0.12 + memtier_benchmark

Protocol: Redis - Clients: 50 - Set To Get Ratio: 10:1

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 50 - Set To Get Ratio: 10:12 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 2849.14, N = 31331731.411. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

Redis 7.0.12 + memtier_benchmark

Protocol: Redis - Clients: 50 - Set To Get Ratio: 1:10

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 50 - Set To Get Ratio: 1:102 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 4611.65, N = 31517459.591. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

Redis 7.0.12 + memtier_benchmark

Protocol: Redis - Clients: 100 - Set To Get Ratio: 10:1

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 100 - Set To Get Ratio: 10:12 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 4259.56, N = 31277786.021. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

Redis 7.0.12 + memtier_benchmark

Protocol: Redis - Clients: 100 - Set To Get Ratio: 1:10

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 100 - Set To Get Ratio: 1:102 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 8411.62, N = 31481597.001. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

John The Ripper

Test: Blowfish

OpenBenchmarking.orgReal C/S, More Is BetterJohn The Ripper 2023.03.14Test: Blowfish2 x Intel Xeon E5-2680 v47K14K21K28K35KSE +/- 66.69, N = 3306611. (CC) gcc options: -m64 -lssl -lcrypto -fopenmp -lgmp -lm -lrt -lz -ldl -lcrypt

Redis

Test: GET - Parallel Connections: 50

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 7.0.4Test: GET - Parallel Connections: 502 x Intel Xeon E5-2680 v4500K1000K1500K2000K2500KSE +/- 14437.10, N = 32246300.921. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Redis

Test: SET - Parallel Connections: 50

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 7.0.4Test: SET - Parallel Connections: 502 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 15716.01, N = 51574262.851. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Redis

Test: GET - Parallel Connections: 500

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 7.0.4Test: GET - Parallel Connections: 5002 x Intel Xeon E5-2680 v4400K800K1200K1600K2000KSE +/- 19833.99, N = 31769970.881. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Redis

Test: LPOP - Parallel Connections: 50

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 7.0.4Test: LPOP - Parallel Connections: 502 x Intel Xeon E5-2680 v4500K1000K1500K2000K2500KSE +/- 17716.01, N = 152203333.821. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Redis

Test: SADD - Parallel Connections: 50

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 7.0.4Test: SADD - Parallel Connections: 502 x Intel Xeon E5-2680 v4400K800K1200K1600K2000KSE +/- 13772.36, N = 91719230.851. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Redis

Test: SET - Parallel Connections: 500

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 7.0.4Test: SET - Parallel Connections: 5002 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 22841.41, N = 121403511.361. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Redis

Test: GET - Parallel Connections: 1000

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 7.0.4Test: GET - Parallel Connections: 10002 x Intel Xeon E5-2680 v4400K800K1200K1600K2000KSE +/- 7775.96, N = 31783987.961. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Redis

Test: LPOP - Parallel Connections: 500

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 7.0.4Test: LPOP - Parallel Connections: 5002 x Intel Xeon E5-2680 v4400K800K1200K1600K2000KSE +/- 30966.88, N = 151955391.541. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Redis

Test: LPUSH - Parallel Connections: 50

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 7.0.4Test: LPUSH - Parallel Connections: 502 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 17610.01, N = 31390421.081. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Redis

Test: SADD - Parallel Connections: 500

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 7.0.4Test: SADD - Parallel Connections: 5002 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 8955.86, N = 31542038.121. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Redis

Test: SET - Parallel Connections: 1000

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 7.0.4Test: SET - Parallel Connections: 10002 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 13555.82, N = 61407706.521. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Redis

Test: LPOP - Parallel Connections: 1000

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 7.0.4Test: LPOP - Parallel Connections: 10002 x Intel Xeon E5-2680 v4400K800K1200K1600K2000KSE +/- 116725.42, N = 121785408.101. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Redis

Test: LPUSH - Parallel Connections: 500

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 7.0.4Test: LPUSH - Parallel Connections: 5002 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 10915.30, N = 71251818.181. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Redis

Test: SADD - Parallel Connections: 1000

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 7.0.4Test: SADD - Parallel Connections: 10002 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 10619.90, N = 31571264.331. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Redis

Test: LPUSH - Parallel Connections: 1000

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 7.0.4Test: LPUSH - Parallel Connections: 10002 x Intel Xeon E5-2680 v4300K600K900K1200K1500KSE +/- 7237.07, N = 31236129.421. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Numpy Benchmark

OpenBenchmarking.orgScore, More Is BetterNumpy Benchmark2 x Intel Xeon E5-2680 v460120180240300SE +/- 0.46, N = 3270.37

CLOMP

Static OMP Speedup

OpenBenchmarking.orgSpeedup, More Is BetterCLOMP 1.2Static OMP Speedup2 x Intel Xeon E5-2680 v4612182430SE +/- 0.29, N = 1525.61. (CC) gcc options: -fopenmp -O3 -lm

QuantLib

Size: S

OpenBenchmarking.orgtasks/s, More Is BetterQuantLib 1.35-devSize: S2 x Intel Xeon E5-2680 v43691215SE +/- 0.01, N = 310.441. (CXX) g++ options: -O3 -fPIE -pie

QuantLib

Size: XXS

OpenBenchmarking.orgtasks/s, More Is BetterQuantLib 1.35-devSize: XXS2 x Intel Xeon E5-2680 v43691215SE +/- 0.01, N = 310.951. (CXX) g++ options: -O3 -fPIE -pie

Llama.cpp

Model: Meta-Llama-3-8B-Instruct-Q8_0.gguf

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b3067Model: Meta-Llama-3-8B-Instruct-Q8_0.gguf2 x Intel Xeon E5-2680 v41.16332.32663.48994.65325.8165SE +/- 0.13, N = 125.171. (CXX) g++ options: -std=c++11 -fPIC -O3 -pthread -march=native -mtune=native -lopenblas

Llamafile

Test: mistral-7b-instruct-v0.2.Q5_K_M - Acceleration: CPU

OpenBenchmarking.orgTokens Per Second, More Is BetterLlamafile 0.8.6Test: mistral-7b-instruct-v0.2.Q5_K_M - Acceleration: CPU2 x Intel Xeon E5-2680 v41.3052.613.9155.226.525SE +/- 0.17, N = 125.80

Llama.cpp

Model: llama-2-7b.Q4_0.gguf

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b1808Model: llama-2-7b.Q4_0.gguf2 x Intel Xeon E5-2680 v4246810SE +/- 0.04, N = 37.481. (CXX) g++ options: -std=c++11 -fPIC -O3 -pthread -march=native -mtune=native -lopenblas

Llama.cpp

Model: llama-2-13b.Q4_0.gguf

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b1808Model: llama-2-13b.Q4_0.gguf2 x Intel Xeon E5-2680 v4246810SE +/- 0.24, N = 126.411. (CXX) g++ options: -std=c++11 -fPIC -O3 -pthread -march=native -mtune=native -lopenblas

Llama.cpp

Model: llama-2-70b-chat.Q5_0.gguf

OpenBenchmarking.orgTokens Per Second, More Is BetterLlama.cpp b1808Model: llama-2-70b-chat.Q5_0.gguf2 x Intel Xeon E5-2680 v40.20250.4050.60750.811.0125SE +/- 0.03, N = 90.901. (CXX) g++ options: -std=c++11 -fPIC -O3 -pthread -march=native -mtune=native -lopenblas

NAS Parallel Benchmarks

Test / Class: EP.C

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: EP.C2 x Intel Xeon E5-2680 v46001200180024003000SE +/- 10.98, N = 32736.221. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.4

NAS Parallel Benchmarks

Test / Class: LU.C

OpenBenchmarking.orgTotal Mop/s, More Is BetterNAS Parallel Benchmarks 3.4Test / Class: LU.C2 x Intel Xeon E5-2680 v410K20K30K40K50KSE +/- 40.59, N = 346814.611. (F9X) gfortran options: -O3 -march=native -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz 2. Open MPI 4.1.4

Apache Siege

Concurrent Users: 10

OpenBenchmarking.orgTransactions Per Second, More Is BetterApache Siege 2.4.62Concurrent Users: 102 x Intel Xeon E5-2680 v45K10K15K20K25KSE +/- 129.08, N = 323642.071. (CC) gcc options: -O2 -lpthread -ldl -lssl -lcrypto -lz

Apache Siege

Concurrent Users: 50

OpenBenchmarking.orgTransactions Per Second, More Is BetterApache Siege 2.4.62Concurrent Users: 502 x Intel Xeon E5-2680 v45K10K15K20K25KSE +/- 8.83, N = 322369.701. (CC) gcc options: -O2 -lpthread -ldl -lssl -lcrypto -lz

Apache Siege

Concurrent Users: 100

OpenBenchmarking.orgTransactions Per Second, More Is BetterApache Siege 2.4.62Concurrent Users: 1002 x Intel Xeon E5-2680 v45K10K15K20K25KSE +/- 38.51, N = 321617.091. (CC) gcc options: -O2 -lpthread -ldl -lssl -lcrypto -lz

Apache Siege

Concurrent Users: 200

OpenBenchmarking.orgTransactions Per Second, More Is BetterApache Siege 2.4.62Concurrent Users: 2002 x Intel Xeon E5-2680 v44K8K12K16K20KSE +/- 52.17, N = 320413.981. (CC) gcc options: -O2 -lpthread -ldl -lssl -lcrypto -lz

Apache Siege

Concurrent Users: 500

OpenBenchmarking.orgTransactions Per Second, More Is BetterApache Siege 2.4.62Concurrent Users: 5002 x Intel Xeon E5-2680 v44K8K12K16K20KSE +/- 71.12, N = 320425.821. (CC) gcc options: -O2 -lpthread -ldl -lssl -lcrypto -lz

Apache Siege

Concurrent Users: 1000

OpenBenchmarking.orgTransactions Per Second, More Is BetterApache Siege 2.4.62Concurrent Users: 10002 x Intel Xeon E5-2680 v44K8K12K16K20KSE +/- 81.59, N = 320260.731. (CC) gcc options: -O2 -lpthread -ldl -lssl -lcrypto -lz

InfluxDB

Concurrent Streams: 4 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 10000

OpenBenchmarking.orgval/sec, More Is BetterInfluxDB 1.8.2Concurrent Streams: 4 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 100002 x Intel Xeon E5-2680 v460K120K180K240K300KSE +/- 3039.46, N = 3281287.4

InfluxDB

Concurrent Streams: 64 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 10000

OpenBenchmarking.orgval/sec, More Is BetterInfluxDB 1.8.2Concurrent Streams: 64 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 100002 x Intel Xeon E5-2680 v4200K400K600K800K1000KSE +/- 6591.71, N = 12938427.0

InfluxDB

Concurrent Streams: 1024 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 10000

OpenBenchmarking.orgval/sec, More Is BetterInfluxDB 1.8.2Concurrent Streams: 1024 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 100002 x Intel Xeon E5-2680 v4200K400K600K800K1000KSE +/- 2360.15, N = 3966559.2

ctx_clock

Context Switch Time

OpenBenchmarking.orgClocks, Fewer Is Betterctx_clockContext Switch Time2 x Intel Xeon E5-2680 v42004006008001000SE +/- 0.67, N = 31063

Pennant

Test: sedovbig

OpenBenchmarking.orgHydro Cycle Time - Seconds, Fewer Is BetterPennant 1.0.1Test: sedovbig2 x Intel Xeon E5-2680 v4918273645SE +/- 0.05, N = 338.221. (CXX) g++ options: -fopenmp -lmpi_cxx -lmpi

Pennant

Test: leblancbig

OpenBenchmarking.orgHydro Cycle Time - Seconds, Fewer Is BetterPennant 1.0.1Test: leblancbig2 x Intel Xeon E5-2680 v4510152025SE +/- 0.04, N = 321.821. (CXX) g++ options: -fopenmp -lmpi_cxx -lmpi

TensorFlow Lite

Model: SqueezeNet

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: SqueezeNet2 x Intel Xeon E5-2680 v49001800270036004500SE +/- 67.64, N = 123967.06

TensorFlow Lite

Model: Inception V4

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: Inception V42 x Intel Xeon E5-2680 v47K14K21K28K35KSE +/- 319.16, N = 1534692.4

TensorFlow Lite

Model: NASNet Mobile

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: NASNet Mobile2 x Intel Xeon E5-2680 v49K18K27K36K45KSE +/- 722.59, N = 1541794.0

TensorFlow Lite

Model: Mobilenet Float

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: Mobilenet Float2 x Intel Xeon E5-2680 v45001000150020002500SE +/- 35.88, N = 122491.03

TensorFlow Lite

Model: Mobilenet Quant

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: Mobilenet Quant2 x Intel Xeon E5-2680 v48001600240032004000SE +/- 29.07, N = 33904.20

TensorFlow Lite

Model: Inception ResNet V2

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: Inception ResNet V22 x Intel Xeon E5-2680 v49K18K27K36K45KSE +/- 707.81, N = 1542979.1

PyBench

Total For Average Test Times

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyBench 2018-02-16Total For Average Test Times2 x Intel Xeon E5-2680 v42004006008001000SE +/- 1.20, N = 31161

Renaissance

Test: Savina Reactors.IO

OpenBenchmarking.orgms, Fewer Is BetterRenaissance 0.14Test: Savina Reactors.IO2 x Intel Xeon E5-2680 v42K4K6K8K10KSE +/- 136.43, N = 1211385.6MIN: 10698.19 / MAX: 20243.26

oneDNN

Harness: IP Shapes 1D - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.6Harness: IP Shapes 1D - Engine: CPU2 x Intel Xeon E5-2680 v40.75681.51362.27043.02723.784SE +/- 0.00512, N = 33.36355MIN: 3.181. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -fcf-protection=full -pie -ldl

oneDNN

Harness: IP Shapes 3D - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.6Harness: IP Shapes 3D - Engine: CPU2 x Intel Xeon E5-2680 v43691215SE +/- 0.03, N = 312.04MIN: 11.851. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -fcf-protection=full -pie -ldl

oneDNN

Harness: Convolution Batch Shapes Auto - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.6Harness: Convolution Batch Shapes Auto - Engine: CPU2 x Intel Xeon E5-2680 v448121620SE +/- 0.02, N = 314.00MIN: 13.811. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -fcf-protection=full -pie -ldl

oneDNN

Harness: Deconvolution Batch shapes_1d - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.6Harness: Deconvolution Batch shapes_1d - Engine: CPU2 x Intel Xeon E5-2680 v43691215SE +/- 0.13, N = 313.17MIN: 9.291. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -fcf-protection=full -pie -ldl

oneDNN

Harness: Deconvolution Batch shapes_3d - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.6Harness: Deconvolution Batch shapes_3d - Engine: CPU2 x Intel Xeon E5-2680 v41.17382.34763.52144.69525.869SE +/- 0.04255, N = 155.21706MIN: 4.911. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -fcf-protection=full -pie -ldl

oneDNN

Harness: Recurrent Neural Network Training - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.6Harness: Recurrent Neural Network Training - Engine: CPU2 x Intel Xeon E5-2680 v4400800120016002000SE +/- 10.64, N = 31845.75MIN: 1764.71. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -fcf-protection=full -pie -ldl

oneDNN

Harness: Recurrent Neural Network Inference - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.6Harness: Recurrent Neural Network Inference - Engine: CPU2 x Intel Xeon E5-2680 v42004006008001000SE +/- 7.40, N = 31018.67MIN: 952.11. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -fcf-protection=full -pie -ldl

Mobile Neural Network

Model: nasnet

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: nasnet2 x Intel Xeon E5-2680 v4510152025SE +/- 0.35, N = 922.03MIN: 16.48 / MAX: 47.881. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

Mobile Neural Network

Model: mobilenetV3

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: mobilenetV32 x Intel Xeon E5-2680 v40.78841.57682.36523.15363.942SE +/- 0.036, N = 93.504MIN: 2.93 / MAX: 5.551. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

Mobile Neural Network

Model: squeezenetv1.1

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: squeezenetv1.12 x Intel Xeon E5-2680 v41.31062.62123.93185.24246.553SE +/- 0.151, N = 95.825MIN: 4.27 / MAX: 16.911. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

Mobile Neural Network

Model: resnet-v2-50

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: resnet-v2-502 x Intel Xeon E5-2680 v4612182430SE +/- 0.18, N = 926.96MIN: 24.27 / MAX: 169.421. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

Mobile Neural Network

Model: SqueezeNetV1.0

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: SqueezeNetV1.02 x Intel Xeon E5-2680 v4246810SE +/- 0.154, N = 97.976MIN: 6.03 / MAX: 23.011. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

Mobile Neural Network

Model: MobileNetV2_224

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: MobileNetV2_2242 x Intel Xeon E5-2680 v41.11312.22623.33934.45245.5655SE +/- 0.071, N = 94.947MIN: 3.92 / MAX: 8.481. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

Mobile Neural Network

Model: mobilenet-v1-1.0

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: mobilenet-v1-1.02 x Intel Xeon E5-2680 v40.84871.69742.54613.39484.2435SE +/- 0.024, N = 93.772MIN: 3.26 / MAX: 10.661. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

Mobile Neural Network

Model: inception-v3

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: inception-v32 x Intel Xeon E5-2680 v4918273645SE +/- 0.24, N = 938.38MIN: 29.94 / MAX: 109.161. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

NCNN

Target: CPU - Model: mobilenet

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: mobilenet2 x Intel Xeon E5-2680 v4510152025SE +/- 0.27, N = 1219.56MIN: 17.72 / MAX: 164.141. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU-v2-v2 - Model: mobilenet-v2

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU-v2-v2 - Model: mobilenet-v22 x Intel Xeon E5-2680 v4246810SE +/- 0.24, N = 127.93MIN: 7.1 / MAX: 581.191. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU-v3-v3 - Model: mobilenet-v3

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU-v3-v3 - Model: mobilenet-v32 x Intel Xeon E5-2680 v4246810SE +/- 0.08, N = 127.49MIN: 6.65 / MAX: 69.831. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU - Model: shufflenet-v2

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: shufflenet-v22 x Intel Xeon E5-2680 v4246810SE +/- 0.11, N = 128.79MIN: 7.42 / MAX: 103.671. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU - Model: mnasnet

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: mnasnet2 x Intel Xeon E5-2680 v4246810SE +/- 0.07, N = 127.06MIN: 5.98 / MAX: 52.181. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU - Model: efficientnet-b0

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: efficientnet-b02 x Intel Xeon E5-2680 v43691215SE +/- 0.11, N = 1211.22MIN: 10.09 / MAX: 196.481. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU - Model: blazeface

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: blazeface2 x Intel Xeon E5-2680 v40.8461.6922.5383.3844.23SE +/- 0.06, N = 123.76MIN: 3.22 / MAX: 97.741. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU - Model: googlenet

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: googlenet2 x Intel Xeon E5-2680 v4510152025SE +/- 0.14, N = 1218.36MIN: 16.63 / MAX: 156.481. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU - Model: vgg16

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: vgg162 x Intel Xeon E5-2680 v4918273645SE +/- 0.41, N = 1241.46MIN: 36.72 / MAX: 331.431. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU - Model: resnet18

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: resnet182 x Intel Xeon E5-2680 v43691215SE +/- 0.08, N = 1211.58MIN: 10.7 / MAX: 110.311. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU - Model: alexnet

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: alexnet2 x Intel Xeon E5-2680 v43691215SE +/- 0.10, N = 129.48MIN: 8.34 / MAX: 111.391. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU - Model: resnet50

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: resnet502 x Intel Xeon E5-2680 v4612182430SE +/- 0.35, N = 1223.76MIN: 20.94 / MAX: 184.921. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPUv2-yolov3v2-yolov3 - Model: mobilenetv2-yolov3

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPUv2-yolov3v2-yolov3 - Model: mobilenetv2-yolov32 x Intel Xeon E5-2680 v4510152025SE +/- 0.27, N = 1219.56MIN: 17.72 / MAX: 164.141. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU - Model: yolov4-tiny

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: yolov4-tiny2 x Intel Xeon E5-2680 v4714212835SE +/- 0.25, N = 1232.22MIN: 29.25 / MAX: 192.051. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU - Model: squeezenet_ssd

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: squeezenet_ssd2 x Intel Xeon E5-2680 v4510152025SE +/- 0.19, N = 1219.85MIN: 16.83 / MAX: 168.171. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU - Model: regnety_400m

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: regnety_400m2 x Intel Xeon E5-2680 v4816243240SE +/- 0.27, N = 1233.52MIN: 30.18 / MAX: 387.551. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU - Model: vision_transformer

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: vision_transformer2 x Intel Xeon E5-2680 v420406080100SE +/- 0.25, N = 1282.40MIN: 76.17 / MAX: 439.371. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: CPU - Model: FastestDet

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: FastestDet2 x Intel Xeon E5-2680 v43691215SE +/- 0.12, N = 1210.09MIN: 8.51 / MAX: 179.651. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU - Model: mobilenet

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: mobilenet2 x Intel Xeon E5-2680 v4510152025SE +/- 0.27, N = 319.60MIN: 18.34 / MAX: 116.671. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU-v2-v2 - Model: mobilenet-v2

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU-v2-v2 - Model: mobilenet-v22 x Intel Xeon E5-2680 v4246810SE +/- 0.20, N = 37.67MIN: 7.11 / MAX: 73.781. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU-v3-v3 - Model: mobilenet-v3

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU-v3-v3 - Model: mobilenet-v32 x Intel Xeon E5-2680 v4246810SE +/- 0.37, N = 37.68MIN: 6.95 / MAX: 135.31. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU - Model: shufflenet-v2

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: shufflenet-v22 x Intel Xeon E5-2680 v4246810SE +/- 0.12, N = 38.66MIN: 7.95 / MAX: 54.971. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU - Model: mnasnet

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: mnasnet2 x Intel Xeon E5-2680 v4246810SE +/- 0.19, N = 37.07MIN: 6.55 / MAX: 55.911. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU - Model: efficientnet-b0

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: efficientnet-b02 x Intel Xeon E5-2680 v43691215SE +/- 0.13, N = 311.31MIN: 10.54 / MAX: 129.331. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU - Model: blazeface

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: blazeface2 x Intel Xeon E5-2680 v40.87751.7552.63253.514.3875SE +/- 0.19, N = 33.90MIN: 3.45 / MAX: 32.381. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU - Model: googlenet

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: googlenet2 x Intel Xeon E5-2680 v4510152025SE +/- 0.09, N = 318.75MIN: 17.58 / MAX: 61.731. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU - Model: vgg16

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: vgg162 x Intel Xeon E5-2680 v41020304050SE +/- 0.15, N = 342.45MIN: 38.84 / MAX: 131.471. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU - Model: resnet18

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: resnet182 x Intel Xeon E5-2680 v43691215SE +/- 0.21, N = 311.47MIN: 10.74 / MAX: 65.891. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU - Model: alexnet

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: alexnet2 x Intel Xeon E5-2680 v43691215SE +/- 0.26, N = 39.81MIN: 9.08 / MAX: 107.571. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU - Model: resnet50

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: resnet502 x Intel Xeon E5-2680 v4612182430SE +/- 0.32, N = 324.06MIN: 22.21 / MAX: 124.51. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPUv2-yolov3v2-yolov3 - Model: mobilenetv2-yolov3

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPUv2-yolov3v2-yolov3 - Model: mobilenetv2-yolov32 x Intel Xeon E5-2680 v4510152025SE +/- 0.27, N = 319.60MIN: 18.34 / MAX: 116.671. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU - Model: yolov4-tiny

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: yolov4-tiny2 x Intel Xeon E5-2680 v4714212835SE +/- 0.18, N = 332.07MIN: 29.66 / MAX: 113.511. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU - Model: squeezenet_ssd

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: squeezenet_ssd2 x Intel Xeon E5-2680 v4510152025SE +/- 0.44, N = 320.46MIN: 18.81 / MAX: 75.681. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU - Model: regnety_400m

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: regnety_400m2 x Intel Xeon E5-2680 v4816243240SE +/- 0.16, N = 333.70MIN: 31.52 / MAX: 136.781. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU - Model: vision_transformer

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: vision_transformer2 x Intel Xeon E5-2680 v420406080100SE +/- 0.29, N = 382.07MIN: 77.72 / MAX: 211.051. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

NCNN

Target: Vulkan GPU - Model: FastestDet

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: FastestDet2 x Intel Xeon E5-2680 v43691215SE +/- 0.10, N = 39.56MIN: 8.52 / MAX: 49.421. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread

TNN

Target: CPU - Model: DenseNet

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: DenseNet2 x Intel Xeon E5-2680 v48001600240032004000SE +/- 13.71, N = 33758.32MIN: 3547 / MAX: 3972.771. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl

TNN

Target: CPU - Model: MobileNet v2

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: MobileNet v22 x Intel Xeon E5-2680 v480160240320400SE +/- 1.50, N = 3386.78MIN: 377.08 / MAX: 410.11. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl

TNN

Target: CPU - Model: SqueezeNet v2

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: SqueezeNet v22 x Intel Xeon E5-2680 v420406080100SE +/- 0.05, N = 393.73MIN: 92.89 / MAX: 102.531. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl

TNN

Target: CPU - Model: SqueezeNet v1.1

OpenBenchmarking.orgms, Fewer Is BetterTNN 0.3Target: CPU - Model: SqueezeNet v1.12 x Intel Xeon E5-2680 v480160240320400SE +/- 0.19, N = 3351.57MIN: 348.06 / MAX: 361.441. (CXX) g++ options: -fopenmp -pthread -fvisibility=hidden -fvisibility=default -O3 -rdynamic -ldl

OpenVINO

Model: Face Detection FP16 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Face Detection FP16 - Device: CPU2 x Intel Xeon E5-2680 v4400800120016002000SE +/- 0.53, N = 31675.23MIN: 1524.28 / MAX: 1770.41. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Person Detection FP16 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Person Detection FP16 - Device: CPU2 x Intel Xeon E5-2680 v44080120160200SE +/- 0.99, N = 3182.19MIN: 143.44 / MAX: 226.911. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Person Detection FP32 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Person Detection FP32 - Device: CPU2 x Intel Xeon E5-2680 v44080120160200SE +/- 1.19, N = 3179.42MIN: 153.31 / MAX: 227.771. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Vehicle Detection FP16 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Vehicle Detection FP16 - Device: CPU2 x Intel Xeon E5-2680 v4612182430SE +/- 0.32, N = 326.15MIN: 19.52 / MAX: 45.161. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Face Detection FP16-INT8 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Face Detection FP16-INT8 - Device: CPU2 x Intel Xeon E5-2680 v430060090012001500SE +/- 0.25, N = 31352.53MIN: 1293.5 / MAX: 1386.351. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Face Detection Retail FP16 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Face Detection Retail FP16 - Device: CPU2 x Intel Xeon E5-2680 v4246810SE +/- 0.02, N = 36.50MIN: 5.8 / MAX: 20.011. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Road Segmentation ADAS FP16 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Road Segmentation ADAS FP16 - Device: CPU2 x Intel Xeon E5-2680 v41530456075SE +/- 0.38, N = 368.76MIN: 43.42 / MAX: 111.071. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Vehicle Detection FP16-INT8 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Vehicle Detection FP16-INT8 - Device: CPU2 x Intel Xeon E5-2680 v4510152025SE +/- 0.01, N = 320.66MIN: 20.17 / MAX: 30.511. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Weld Porosity Detection FP16 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Weld Porosity Detection FP16 - Device: CPU2 x Intel Xeon E5-2680 v4510152025SE +/- 0.00, N = 318.55MIN: 17.41 / MAX: 36.641. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Face Detection Retail FP16-INT8 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Face Detection Retail FP16-INT8 - Device: CPU2 x Intel Xeon E5-2680 v4246810SE +/- 0.00, N = 36.91MIN: 6.78 / MAX: 13.021. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Road Segmentation ADAS FP16-INT8 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Road Segmentation ADAS FP16-INT8 - Device: CPU2 x Intel Xeon E5-2680 v41020304050SE +/- 0.06, N = 344.46MIN: 42.03 / MAX: 58.711. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Machine Translation EN To DE FP16 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Machine Translation EN To DE FP16 - Device: CPU2 x Intel Xeon E5-2680 v44080120160200SE +/- 0.63, N = 3162.59MIN: 131.74 / MAX: 222.031. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Weld Porosity Detection FP16-INT8 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Weld Porosity Detection FP16-INT8 - Device: CPU2 x Intel Xeon E5-2680 v4918273645SE +/- 0.03, N = 340.46MIN: 39.84 / MAX: 47.661. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Person Vehicle Bike Detection FP16 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Person Vehicle Bike Detection FP16 - Device: CPU2 x Intel Xeon E5-2680 v4510152025SE +/- 0.00, N = 321.23MIN: 17.51 / MAX: 33.531. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Noise Suppression Poconet-Like FP16 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Noise Suppression Poconet-Like FP16 - Device: CPU2 x Intel Xeon E5-2680 v448121620SE +/- 0.01, N = 315.74MIN: 13.82 / MAX: 29.981. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Handwritten English Recognition FP16 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Handwritten English Recognition FP16 - Device: CPU2 x Intel Xeon E5-2680 v41224364860SE +/- 0.04, N = 353.49MIN: 49.19 / MAX: 73.081. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Person Re-Identification Retail FP16 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Person Re-Identification Retail FP16 - Device: CPU2 x Intel Xeon E5-2680 v43691215SE +/- 0.01, N = 312.84MIN: 12.15 / MAX: 23.491. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Age Gender Recognition Retail 0013 FP16 - Device: CPU2 x Intel Xeon E5-2680 v40.3960.7921.1881.5841.98SE +/- 0.01, N = 31.76MIN: 1.68 / MAX: 9.161. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Handwritten English Recognition FP16-INT8 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Handwritten English Recognition FP16-INT8 - Device: CPU2 x Intel Xeon E5-2680 v4306090120150SE +/- 0.36, N = 3130.16MIN: 116.02 / MAX: 172.781. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenVINO

Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2024.0Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPU2 x Intel Xeon E5-2680 v40.29250.5850.87751.171.4625SE +/- 0.00, N = 31.30MIN: 1.27 / MAX: 10.771. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -shared -ldl

OpenCV

Test: DNN - Deep Neural Network

OpenBenchmarking.orgms, Fewer Is BetterOpenCV 4.7Test: DNN - Deep Neural Network2 x Intel Xeon E5-2680 v411K22K33K44K55KSE +/- 2440.79, N = 15505671. (CXX) g++ options: -fPIC -fsigned-char -pthread -fomit-frame-pointer -ffunction-sections -fdata-sections -msse -msse2 -msse3 -fvisibility=hidden -O3 -shared

Neural Magic DeepSparse

Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v42004006008001000SE +/- 2.74, N = 3806.62

Neural Magic DeepSparse

Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v4306090120150SE +/- 0.16, N = 3121.77

Neural Magic DeepSparse

Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v4918273645SE +/- 0.07, N = 340.48

Neural Magic DeepSparse

Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v43691215SE +/- 0.0282, N = 39.9602

Neural Magic DeepSparse

Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v41632486480SE +/- 0.27, N = 371.85

Neural Magic DeepSparse

Model: ResNet-50, Baseline - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: ResNet-50, Baseline - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v43691215SE +/- 0.01, N = 312.73

Neural Magic DeepSparse

Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v448121620SE +/- 0.03, N = 313.68

Neural Magic DeepSparse

Model: ResNet-50, Sparse INT8 - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: ResNet-50, Sparse INT8 - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v40.53291.06581.59872.13162.6645SE +/- 0.0060, N = 32.3685

Neural Magic DeepSparse

Model: Llama2 Chat 7b Quantized - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: Llama2 Chat 7b Quantized - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v42K4K6K8K10KSE +/- 37.55, N = 39614.47

Neural Magic DeepSparse

Model: Llama2 Chat 7b Quantized - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: Llama2 Chat 7b Quantized - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v470140210280350SE +/- 0.72, N = 3314.78

Neural Magic DeepSparse

Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v41632486480SE +/- 0.33, N = 371.71

Neural Magic DeepSparse

Model: CV Classification, ResNet-50 ImageNet - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: CV Classification, ResNet-50 ImageNet - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v43691215SE +/- 0.02, N = 312.78

Neural Magic DeepSparse

Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v4306090120150SE +/- 0.32, N = 3146.34

Neural Magic DeepSparse

Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v4612182430SE +/- 0.06, N = 324.04

Neural Magic DeepSparse

Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v420406080100SE +/- 0.23, N = 393.16

Neural Magic DeepSparse

Model: NLP Text Classification, DistilBERT mnli - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: NLP Text Classification, DistilBERT mnli - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v448121620SE +/- 0.07, N = 317.25

Neural Magic DeepSparse

Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v42004006008001000SE +/- 0.49, N = 3880.92

Neural Magic DeepSparse

Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v4306090120150SE +/- 0.15, N = 3126.92

Neural Magic DeepSparse

Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v420406080100SE +/- 0.07, N = 381.97

Neural Magic DeepSparse

Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v4510152025SE +/- 0.07, N = 322.25

Neural Magic DeepSparse

Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Stream2 x Intel Xeon E5-2680 v42004006008001000SE +/- 1.38, N = 3805.37

Neural Magic DeepSparse

Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Synchronous Single-Stream

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.7Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Synchronous Single-Stream2 x Intel Xeon E5-2680 v4306090120150SE +/- 0.27, N = 3122.21

DaCapo Benchmark

Java Test: Jython

OpenBenchmarking.orgmsec, Fewer Is BetterDaCapo Benchmark 23.11Java Test: Jython2 x Intel Xeon E5-2680 v42K4K6K8K10KSE +/- 47.04, N = 38144

DaCapo Benchmark

Java Test: Tradebeans

OpenBenchmarking.orgmsec, Fewer Is BetterDaCapo Benchmark 23.11Java Test: Tradebeans2 x Intel Xeon E5-2680 v44K8K12K16K20KSE +/- 185.70, N = 319157

Glibc Benchmarks

Benchmark: cos

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: cos2 x Intel Xeon E5-2680 v41530456075SE +/- 0.33, N = 365.511. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Glibc Benchmarks

Benchmark: exp

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: exp2 x Intel Xeon E5-2680 v448121620SE +/- 0.17, N = 315.281. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Glibc Benchmarks

Benchmark: ffs

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: ffs2 x Intel Xeon E5-2680 v40.98721.97442.96163.94884.936SE +/- 0.00101, N = 34.387701. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Glibc Benchmarks

Benchmark: pow

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: pow2 x Intel Xeon E5-2680 v4816243240SE +/- 0.11, N = 334.461. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Glibc Benchmarks

Benchmark: sin

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: sin2 x Intel Xeon E5-2680 v41326395265SE +/- 0.12, N = 355.721. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Glibc Benchmarks

Benchmark: log2

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: log22 x Intel Xeon E5-2680 v43691215SE +/- 0.00, N = 313.161. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Glibc Benchmarks

Benchmark: modf

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: modf2 x Intel Xeon E5-2680 v4246810SE +/- 0.00207, N = 36.581741. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Glibc Benchmarks

Benchmark: sinh

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: sinh2 x Intel Xeon E5-2680 v4612182430SE +/- 0.03, N = 323.941. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Glibc Benchmarks

Benchmark: sqrt

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: sqrt2 x Intel Xeon E5-2680 v41.31682.63363.95045.26726.584SE +/- 0.00191, N = 35.852461. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Glibc Benchmarks

Benchmark: tanh

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: tanh2 x Intel Xeon E5-2680 v4714212835SE +/- 0.02, N = 332.211. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Glibc Benchmarks

Benchmark: asinh

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: asinh2 x Intel Xeon E5-2680 v4714212835SE +/- 0.05, N = 329.421. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Glibc Benchmarks

Benchmark: atanh

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: atanh2 x Intel Xeon E5-2680 v4714212835SE +/- 0.07, N = 331.981. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Glibc Benchmarks

Benchmark: ffsll

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: ffsll2 x Intel Xeon E5-2680 v40.98691.97382.96073.94764.9345SE +/- 0.00074, N = 34.386121. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Glibc Benchmarks

Benchmark: sincos

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: sincos2 x Intel Xeon E5-2680 v4918273645SE +/- 0.34, N = 839.181. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Glibc Benchmarks

Benchmark: pthread_once

OpenBenchmarking.orgns, Fewer Is BetterGlibc Benchmarks 2.39Benchmark: pthread_once2 x Intel Xeon E5-2680 v41.15262.30523.45784.61045.763SE +/- 0.01136, N = 35.122721. (CC) gcc options: -pie -nostdlib -nostartfiles -lgcc -lgcc_s

Rodinia

Test: OpenMP LavaMD

OpenBenchmarking.orgSeconds, Fewer Is BetterRodinia 3.1Test: OpenMP LavaMD2 x Intel Xeon E5-2680 v4306090120150SE +/- 0.28, N = 3133.481. (CXX) g++ options: -O2 -lOpenCL

Timed GCC Compilation

Time To Compile

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed GCC Compilation 13.2Time To Compile2 x Intel Xeon E5-2680 v430060090012001500SE +/- 15.50, N = 51491.16

Timed PHP Compilation

Time To Compile

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed PHP Compilation 8.3.4Time To Compile2 x Intel Xeon E5-2680 v420406080100SE +/- 0.45, N = 378.38

C-Ray

Total Time - 4K, 16 Rays Per Pixel

OpenBenchmarking.orgSeconds, Fewer Is BetterC-Ray 2.0Total Time - 4K, 16 Rays Per Pixel2 x Intel Xeon E5-2680 v40.13910.27820.41730.55640.6955SE +/- 0.005, N = 30.6181. (CC) gcc options: -lpthread -lm

POV-Ray

Trace Time

OpenBenchmarking.orgSeconds, Fewer Is BetterPOV-Ray 3.7.0.7Trace Time2 x Intel Xeon E5-2680 v4612182430SE +/- 0.08, N = 323.761. (CXX) g++ options: -pipe -O3 -ffast-math -march=native -R/usr/lib -lSM -lICE -lX11 -ltiff -ljpeg -lpng -lz -lrt -lm -lboost_thread -lboost_system

Rust Mandelbrot

Time To Complete Serial/Parallel Mandelbrot

OpenBenchmarking.orgSeconds, Fewer Is BetterRust MandelbrotTime To Complete Serial/Parallel Mandelbrot2 x Intel Xeon E5-2680 v41020304050SE +/- 0.13, N = 345.521. (CC) gcc options: -m64 -lgcc_s -lutil -lrt -lpthread -lm -ldl -lc -pie -nodefaultlibs

Cython Benchmark

Test: N-Queens

OpenBenchmarking.orgSeconds, Fewer Is BetterCython Benchmark 0.29.21Test: N-Queens2 x Intel Xeon E5-2680 v4816243240SE +/- 0.21, N = 332.49

Hackbench

Count: 32 - Type: Process

OpenBenchmarking.orgSeconds, Fewer Is BetterHackbenchCount: 32 - Type: Process2 x Intel Xeon E5-2680 v41326395265SE +/- 0.07, N = 357.431. (CC) gcc options: -lpthread

m-queens

Time To Solve

OpenBenchmarking.orgSeconds, Fewer Is Betterm-queens 1.2Time To Solve2 x Intel Xeon E5-2680 v4816243240SE +/- 0.01, N = 336.861. (CXX) g++ options: -fopenmp -O2 -march=native

Rodinia

Test: OpenMP HotSpot3D

OpenBenchmarking.orgSeconds, Fewer Is BetterRodinia 3.1Test: OpenMP HotSpot3D2 x Intel Xeon E5-2680 v4306090120150SE +/- 0.06, N = 3141.431. (CXX) g++ options: -O2 -lOpenCL

Rodinia

Test: OpenMP Leukocyte

OpenBenchmarking.orgSeconds, Fewer Is BetterRodinia 3.1Test: OpenMP Leukocyte2 x Intel Xeon E5-2680 v41530456075SE +/- 0.72, N = 366.191. (CXX) g++ options: -O2 -lOpenCL

Rodinia

Test: OpenMP CFD Solver

OpenBenchmarking.orgSeconds, Fewer Is BetterRodinia 3.1Test: OpenMP CFD Solver2 x Intel Xeon E5-2680 v43691215SE +/- 0.11, N = 311.161. (CXX) g++ options: -O2 -lOpenCL

Rodinia

Test: OpenMP Streamcluster

OpenBenchmarking.orgSeconds, Fewer Is BetterRodinia 3.1Test: OpenMP Streamcluster2 x Intel Xeon E5-2680 v448121620SE +/- 0.13, N = 1513.751. (CXX) g++ options: -O2 -lOpenCL

Xcompact3d Incompact3d

Input: X3D-benchmarking input.i3d

OpenBenchmarking.orgSeconds, Fewer Is BetterXcompact3d Incompact3d 2021-03-11Input: X3D-benchmarking input.i3d2 x Intel Xeon E5-2680 v430060090012001500SE +/- 16.45, N = 91163.961. (F9X) gfortran options: -cpp -O2 -funroll-loops -floop-optimize -fcray-pointer -fbacktrace -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz

Xcompact3d Incompact3d

Input: input.i3d 129 Cells Per Direction

OpenBenchmarking.orgSeconds, Fewer Is BetterXcompact3d Incompact3d 2021-03-11Input: input.i3d 129 Cells Per Direction2 x Intel Xeon E5-2680 v43691215SE +/- 0.10, N = 610.651. (F9X) gfortran options: -cpp -O2 -funroll-loops -floop-optimize -fcray-pointer -fbacktrace -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz

Xcompact3d Incompact3d

Input: input.i3d 193 Cells Per Direction

OpenBenchmarking.orgSeconds, Fewer Is BetterXcompact3d Incompact3d 2021-03-11Input: input.i3d 193 Cells Per Direction2 x Intel Xeon E5-2680 v41122334455SE +/- 0.11, N = 346.691. (F9X) gfortran options: -cpp -O2 -funroll-loops -floop-optimize -fcray-pointer -fbacktrace -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz

OpenRadioss

Model: Bumper Beam

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2023.09.15Model: Bumper Beam2 x Intel Xeon E5-2680 v4306090120150SE +/- 1.00, N = 3143.24

OpenRadioss

Model: Cell Phone Drop Test

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2023.09.15Model: Cell Phone Drop Test2 x Intel Xeon E5-2680 v420406080100SE +/- 0.19, N = 375.12

OpenRadioss

Model: Bird Strike on Windshield

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenRadioss 2023.09.15Model: Bird Strike on Windshield2 x Intel Xeon E5-2680 v450100150200250SE +/- 0.88, N = 3230.48

Blender

Blend File: BMW27 - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.2Blend File: BMW27 - Compute: CPU-Only2 x Intel Xeon E5-2680 v41632486480SE +/- 0.32, N = 372.25

Blender

Blend File: Junkshop - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.2Blend File: Junkshop - Compute: CPU-Only2 x Intel Xeon E5-2680 v420406080100SE +/- 0.16, N = 399.22

Blender

Blend File: Classroom - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.2Blend File: Classroom - Compute: CPU-Only2 x Intel Xeon E5-2680 v450100150200250SE +/- 2.43, N = 4215.31

Blender

Blend File: Fishy Cat - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.2Blend File: Fishy Cat - Compute: CPU-Only2 x Intel Xeon E5-2680 v420406080100SE +/- 0.32, N = 3105.76

Blender

Blend File: Barbershop - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.2Blend File: Barbershop - Compute: CPU-Only2 x Intel Xeon E5-2680 v4160320480640800SE +/- 0.76, N = 3736.44

Blender

Blend File: Pabellon Barcelona - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.2Blend File: Pabellon Barcelona - Compute: CPU-Only2 x Intel Xeon E5-2680 v450100150200250SE +/- 0.28, N = 3233.47

Timed Wasmer Compilation

Time To Compile

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Wasmer Compilation 2.3Time To Compile2 x Intel Xeon E5-2680 v420406080100SE +/- 0.63, N = 376.411. (CC) gcc options: -m64 -ldl -lgcc_s -lutil -lrt -lpthread -lm -lc -pie -nodefaultlibs

Timed Node.js Compilation

Time To Compile

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Node.js Compilation 21.7.2Time To Compile2 x Intel Xeon E5-2680 v4400800120016002000SE +/- 29.50, N = 32052.80

libavif avifenc

Encoder Speed: 0

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 1.0Encoder Speed: 02 x Intel Xeon E5-2680 v4306090120150SE +/- 0.77, N = 3150.331. (CXX) g++ options: -O3 -fPIC -lm

libavif avifenc

Encoder Speed: 2

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 1.0Encoder Speed: 22 x Intel Xeon E5-2680 v420406080100SE +/- 0.28, N = 379.421. (CXX) g++ options: -O3 -fPIC -lm

libavif avifenc

Encoder Speed: 6

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 1.0Encoder Speed: 62 x Intel Xeon E5-2680 v4246810SE +/- 0.043, N = 36.7311. (CXX) g++ options: -O3 -fPIC -lm

libavif avifenc

Encoder Speed: 6, Lossless

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 1.0Encoder Speed: 6, Lossless2 x Intel Xeon E5-2680 v43691215SE +/- 0.07, N = 311.931. (CXX) g++ options: -O3 -fPIC -lm

libavif avifenc

Encoder Speed: 10, Lossless

OpenBenchmarking.orgSeconds, Fewer Is Betterlibavif avifenc 1.0Encoder Speed: 10, Lossless2 x Intel Xeon E5-2680 v4246810SE +/- 0.080, N = 38.3881. (CXX) g++ options: -O3 -fPIC -lm

Timed LLVM Compilation

Build System: Ninja

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed LLVM Compilation 16.0Build System: Ninja2 x Intel Xeon E5-2680 v4110220330440550SE +/- 0.54, N = 3497.46

Timed LLVM Compilation

Build System: Unix Makefiles

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed LLVM Compilation 16.0Build System: Unix Makefiles2 x Intel Xeon E5-2680 v4130260390520650SE +/- 1.83, N = 3604.37

DeepSpeech

Acceleration: CPU

OpenBenchmarking.orgSeconds, Fewer Is BetterDeepSpeech 0.6Acceleration: CPU2 x Intel Xeon E5-2680 v44080120160200SE +/- 0.69, N = 3193.78

R Benchmark

OpenBenchmarking.orgSeconds, Fewer Is BetterR Benchmark2 x Intel Xeon E5-2680 v40.05130.10260.15390.20520.2565SE +/- 0.0016, N = 130.2279

RNNoise

Input: 26 Minute Long Talking Sample

OpenBenchmarking.orgSeconds, Fewer Is BetterRNNoise 0.2Input: 26 Minute Long Talking Sample2 x Intel Xeon E5-2680 v448121620SE +/- 0.26, N = 318.191. (CC) gcc options: -O2 -pedantic -fvisibility=hidden

Numenta Anomaly Benchmark

Detector: KNN CAD

OpenBenchmarking.orgSeconds, Fewer Is BetterNumenta Anomaly Benchmark 1.1Detector: KNN CAD2 x Intel Xeon E5-2680 v44080120160200SE +/- 1.37, N = 3169.28

Numenta Anomaly Benchmark

Detector: Relative Entropy

OpenBenchmarking.orgSeconds, Fewer Is BetterNumenta Anomaly Benchmark 1.1Detector: Relative Entropy2 x Intel Xeon E5-2680 v4510152025SE +/- 0.24, N = 421.62

Numenta Anomaly Benchmark

Detector: Windowed Gaussian

OpenBenchmarking.orgSeconds, Fewer Is BetterNumenta Anomaly Benchmark 1.1Detector: Windowed Gaussian2 x Intel Xeon E5-2680 v4246810SE +/- 0.070, N = 158.543

Numenta Anomaly Benchmark

Detector: Earthgecko Skyline

OpenBenchmarking.orgSeconds, Fewer Is BetterNumenta Anomaly Benchmark 1.1Detector: Earthgecko Skyline2 x Intel Xeon E5-2680 v420406080100SE +/- 0.33, N = 3110.01

Numenta Anomaly Benchmark

Detector: Bayesian Changepoint

OpenBenchmarking.orgSeconds, Fewer Is BetterNumenta Anomaly Benchmark 1.1Detector: Bayesian Changepoint2 x Intel Xeon E5-2680 v41020304050SE +/- 0.55, N = 345.70

Numenta Anomaly Benchmark

Detector: Contextual Anomaly Detector OSE

OpenBenchmarking.orgSeconds, Fewer Is BetterNumenta Anomaly Benchmark 1.1Detector: Contextual Anomaly Detector OSE2 x Intel Xeon E5-2680 v41020304050SE +/- 0.43, N = 343.62

Whisper.cpp

Model: ggml-base.en - Input: 2016 State of the Union

OpenBenchmarking.orgSeconds, Fewer Is BetterWhisper.cpp 1.6.2Model: ggml-base.en - Input: 2016 State of the Union2 x Intel Xeon E5-2680 v44080120160200SE +/- 1.87, N = 3202.231. (CXX) g++ options: -O3 -std=c++11 -fPIC -pthread -msse3 -mssse3 -mavx -mf16c -mfma -mavx2

Whisper.cpp

Model: ggml-small.en - Input: 2016 State of the Union

OpenBenchmarking.orgSeconds, Fewer Is BetterWhisper.cpp 1.6.2Model: ggml-small.en - Input: 2016 State of the Union2 x Intel Xeon E5-2680 v4120240360480600SE +/- 3.14, N = 3543.891. (CXX) g++ options: -O3 -std=c++11 -fPIC -pthread -msse3 -mssse3 -mavx -mf16c -mfma -mavx2

Whisper.cpp

Model: ggml-medium.en - Input: 2016 State of the Union

OpenBenchmarking.orgSeconds, Fewer Is BetterWhisper.cpp 1.6.2Model: ggml-medium.en - Input: 2016 State of the Union2 x Intel Xeon E5-2680 v430060090012001500SE +/- 3.76, N = 31539.311. (CXX) g++ options: -O3 -std=c++11 -fPIC -pthread -msse3 -mssse3 -mavx -mf16c -mfma -mavx2

XNNPACK

Model: FP32MobileNetV1

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK b7b048Model: FP32MobileNetV12 x Intel Xeon E5-2680 v45001000150020002500SE +/- 19.22, N = 322991. (CXX) g++ options: -O3 -lrt -lm

XNNPACK

Model: FP32MobileNetV2

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK b7b048Model: FP32MobileNetV22 x Intel Xeon E5-2680 v48001600240032004000SE +/- 275.46, N = 337251. (CXX) g++ options: -O3 -lrt -lm

XNNPACK

Model: FP32MobileNetV3Large

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK b7b048Model: FP32MobileNetV3Large2 x Intel Xeon E5-2680 v411002200330044005500SE +/- 109.39, N = 349331. (CXX) g++ options: -O3 -lrt -lm

XNNPACK

Model: FP32MobileNetV3Small

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK b7b048Model: FP32MobileNetV3Small2 x Intel Xeon E5-2680 v48001600240032004000SE +/- 120.36, N = 336741. (CXX) g++ options: -O3 -lrt -lm

XNNPACK

Model: FP16MobileNetV1

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK b7b048Model: FP16MobileNetV12 x Intel Xeon E5-2680 v46001200180024003000SE +/- 54.55, N = 326631. (CXX) g++ options: -O3 -lrt -lm

XNNPACK

Model: FP16MobileNetV2

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK b7b048Model: FP16MobileNetV22 x Intel Xeon E5-2680 v48001600240032004000SE +/- 209.77, N = 336851. (CXX) g++ options: -O3 -lrt -lm

XNNPACK

Model: FP16MobileNetV3Large

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK b7b048Model: FP16MobileNetV3Large2 x Intel Xeon E5-2680 v411002200330044005500SE +/- 3.84, N = 351191. (CXX) g++ options: -O3 -lrt -lm

XNNPACK

Model: FP16MobileNetV3Small

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK b7b048Model: FP16MobileNetV3Small2 x Intel Xeon E5-2680 v47001400210028003500SE +/- 44.00, N = 334771. (CXX) g++ options: -O3 -lrt -lm

XNNPACK

Model: QS8MobileNetV2

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK b7b048Model: QS8MobileNetV22 x Intel Xeon E5-2680 v47001400210028003500SE +/- 33.39, N = 332931. (CXX) g++ options: -O3 -lrt -lm


Phoronix Test Suite v10.8.5