8490h april

2 x Intel Xeon Platinum 8490H testing with a Quanta Cloud S6Q-MB-MPS (3A10.uh BIOS) and ASPEED on Ubuntu 22.04 via the Phoronix Test Suite.

HTML result view exported from: https://openbenchmarking.org/result/2304136-NE-8490HAPRI45&gru&sor&rro.

8490h aprilProcessorMotherboardChipsetMemoryDiskGraphicsMonitorNetworkOSKernelDesktopDisplay ServerVulkanCompilerFile-SystemScreen Resolutionabcde2 x Intel Xeon Platinum 8490H @ 3.50GHz (120 Cores / 240 Threads)Quanta Cloud S6Q-MB-MPS (3A10.uh BIOS)Intel Device 1bce16 x 64 GB 4800MT/s Samsung M321R8GA0BB0-CQKEG2 x 1920GB SAMSUNG MZWLJ1T9HBJR-00007 + 960GB INTEL SSDSC2KG96ASPEEDVGA HDMI4 x Intel E810-C for QSFP + 2 x Intel X710 for 10GBASE-TUbuntu 22.046.2.0-060200rc7daily20230208-generic (x86_64)GNOME Shell 42.2X Server 1.21.1.31.2.204GCC 11.3.0 + Clang 14.0.0-1ubuntu1ext41920x1080OpenBenchmarking.orgKernel Details- Transparent Huge Pages: madviseCompiler Details- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-xKiWfi/gcc-11-11.3.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-xKiWfi/gcc-11-11.3.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- Scaling Governor: intel_pstate performance (EPP: performance) - CPU Microcode: 0x2b0000c0Python Details- Python 3.10.6Security Details- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling PBRSB-eIBRS: SW sequence + srbds: Not affected + tsx_async_abort: Not affected

8490h aprilvvenc: Bosphorus 4K - Fastvvenc: Bosphorus 4K - Fastervvenc: Bosphorus 1080p - Fastvvenc: Bosphorus 1080p - Fastertensorflow: CPU - 16 - AlexNettensorflow: CPU - 32 - AlexNettensorflow: CPU - 64 - AlexNettensorflow: CPU - 256 - AlexNettensorflow: CPU - 512 - AlexNettensorflow: CPU - 16 - GoogLeNettensorflow: CPU - 16 - ResNet-50tensorflow: CPU - 32 - GoogLeNettensorflow: CPU - 32 - ResNet-50tensorflow: CPU - 64 - GoogLeNettensorflow: CPU - 64 - ResNet-50tensorflow: CPU - 256 - GoogLeNettensorflow: CPU - 256 - ResNet-50tensorflow: CPU - 512 - GoogLeNettensorflow: CPU - 512 - ResNet-50srsran: Downlink Processor Benchmarksrsran: PUSCH Processor Benchmark, Throughput Totalsrsran: PUSCH Processor Benchmark, Throughput Threadnginx: 500apache: 500onednn: IP Shapes 1D - f32 - CPUonednn: IP Shapes 3D - f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUonednn: IP Shapes 1D - bf16bf16bf16 - CPUonednn: IP Shapes 3D - bf16bf16bf16 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Deconvolution Batch shapes_1d - f32 - CPUonednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - bf16bf16bf16 - CPUonednn: Deconvolution Batch shapes_1d - bf16bf16bf16 - CPUonednn: Deconvolution Batch shapes_3d - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUblender: BMW27 - CPU-Onlyblender: Classroom - CPU-Onlyblender: Fishy Cat - CPU-Onlyblender: Barbershop - CPU-Onlyblender: Pabellon Barcelona - CPU-Onlyabcde6.30810.06517.14728.66372.88531.68743.731091.421227.69173.6464.28257.3383.13348103.48444.17130.44472.26135.88326.57122.429.9250533.3780395.593.567592.497575.193790.8724765.884723.167790.40871114.26670.7244130.239960.4346580.2289711216.99881.2321304.570.2287410.4703360.464269873.1471205.29861.14814.0336.519.36147.2548.816.33210.05517.39630.988386.55556.34741.871077.571231.85185.7864.31267.0284.42342.26102.21442.93128.89465.31134.34324.26898.629.8246156.1183834.813.050002.676994.624780.9784285.387343.046380.40298314.62840.7187460.3144270.4100290.2251971155.77840.9561232.870.2231420.4513930.466045832.3381184.14878.48914.2036.6619.70147.7347.846.3149.96717.24430.211370.67557.68751.671063.471214.36176.8463.78265.0884.98334.11104.52437.97127.52467.33135.22326.76547.429.7246619.5477777.033.635852.528485.307691.153325.420712.913810.40567714.54440.7164190.4335230.3974350.2193411209.39852.5771081.70.217420.4578930.457996844.3581228.77904.26814.2136.7919.94146.5947.656.4439.95617.21127.619391.88536.63739.021071.621225.54184.863.97249.7484.17346.11103.14441.29128.8462.37134.76320.87079.528.8247581.6484694.763.504852.374794.875480.9813614.977182.837540.40841614.48910.7113060.2961520.3919570.2257421182.32731.0951205.380.219490.440410.462589832.5741184.12888.73214.0436.3120.13147.1847.736.38810.06716.78930.369386.34564.79745.331062.061230.3185.2264.96270.3183.45346.2102.87441.44128.23469.14133.9324.16774.528.9248416.8585357.843.446772.808695.347550.9893085.5583.021880.40032514.22120.7122480.3055030.4137350.2193481120.64848.6521200.190.222020.4462320.453885845.7261112.04818.43814.336.3619.54148.1147.43OpenBenchmarking.org

VVenC

Video Input: Bosphorus 4K - Video Preset: Fast

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.8Video Input: Bosphorus 4K - Video Preset: Fastacbed246810SE +/- 0.037, N = 36.3086.3146.3326.3886.4431. (CXX) g++ options: -O3 -flto -fno-fat-lto-objects -flto=auto

VVenC

Video Input: Bosphorus 4K - Video Preset: Faster

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.8Video Input: Bosphorus 4K - Video Preset: Fasterdcbae3691215SE +/- 0.068, N = 139.9569.96710.05510.06510.0671. (CXX) g++ options: -O3 -flto -fno-fat-lto-objects -flto=auto

VVenC

Video Input: Bosphorus 1080p - Video Preset: Fast

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.8Video Input: Bosphorus 1080p - Video Preset: Fasteadcb48121620SE +/- 0.04, N = 316.7917.1517.2117.2417.401. (CXX) g++ options: -O3 -flto -fno-fat-lto-objects -flto=auto

VVenC

Video Input: Bosphorus 1080p - Video Preset: Faster

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.8Video Input: Bosphorus 1080p - Video Preset: Fasterdaceb714212835SE +/- 0.17, N = 327.6228.6630.2130.3730.991. (CXX) g++ options: -O3 -flto -fno-fat-lto-objects -flto=auto

TensorFlow

Device: CPU - Batch Size: 16 - Model: AlexNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 16 - Model: AlexNetcaebd90180270360450SE +/- 3.12, N = 3370.67372.88386.34386.55391.88

TensorFlow

Device: CPU - Batch Size: 32 - Model: AlexNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 32 - Model: AlexNetadbce120240360480600SE +/- 5.22, N = 6531.68536.63556.34557.68564.79

TensorFlow

Device: CPU - Batch Size: 64 - Model: AlexNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 64 - Model: AlexNetdbaec160320480640800SE +/- 6.00, N = 3739.02741.87743.73745.33751.67

TensorFlow

Device: CPU - Batch Size: 256 - Model: AlexNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 256 - Model: AlexNetecdba2004006008001000SE +/- 5.13, N = 31062.061063.471071.621077.571091.42

TensorFlow

Device: CPU - Batch Size: 512 - Model: AlexNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 512 - Model: AlexNetcdaeb30060090012001500SE +/- 2.69, N = 31214.361225.541227.691230.301231.85

TensorFlow

Device: CPU - Batch Size: 16 - Model: GoogLeNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 16 - Model: GoogLeNetacdeb4080120160200SE +/- 1.60, N = 3173.64176.84184.80185.22185.78

TensorFlow

Device: CPU - Batch Size: 16 - Model: ResNet-50

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 16 - Model: ResNet-50cdabe1428425670SE +/- 0.45, N = 363.7863.9764.2864.3164.96

TensorFlow

Device: CPU - Batch Size: 32 - Model: GoogLeNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 32 - Model: GoogLeNetdacbe60120180240300SE +/- 0.97, N = 3249.74257.33265.08267.02270.31

TensorFlow

Device: CPU - Batch Size: 32 - Model: ResNet-50

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 32 - Model: ResNet-50aedbc20406080100SE +/- 0.33, N = 383.1383.4584.1784.4284.98

TensorFlow

Device: CPU - Batch Size: 64 - Model: GoogLeNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 64 - Model: GoogLeNetcbdea80160240320400SE +/- 2.89, N = 3334.11342.26346.11346.20348.00

TensorFlow

Device: CPU - Batch Size: 64 - Model: ResNet-50

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 64 - Model: ResNet-50bedac20406080100SE +/- 0.44, N = 3102.21102.87103.14103.48104.52

TensorFlow

Device: CPU - Batch Size: 256 - Model: GoogLeNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 256 - Model: GoogLeNetcdeba100200300400500SE +/- 3.52, N = 3437.97441.29441.44442.93444.17

TensorFlow

Device: CPU - Batch Size: 256 - Model: ResNet-50

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 256 - Model: ResNet-50cedba306090120150SE +/- 0.05, N = 3127.52128.23128.80128.89130.44

TensorFlow

Device: CPU - Batch Size: 512 - Model: GoogLeNet

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 512 - Model: GoogLeNetdbcea100200300400500SE +/- 4.06, N = 3462.37465.31467.33469.14472.26

TensorFlow

Device: CPU - Batch Size: 512 - Model: ResNet-50

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.12Device: CPU - Batch Size: 512 - Model: ResNet-50ebdca306090120150SE +/- 1.30, N = 3133.90134.34134.76135.22135.88

srsRAN Project

Test: Downlink Processor Benchmark

OpenBenchmarking.orgMbps, More Is BettersrsRAN Project 23.3Test: Downlink Processor Benchmarkdebac70140210280350SE +/- 2.03, N = 3320.8324.1324.2326.5326.7MIN: 71.3 / MAX: 723.1MIN: 69.9 / MAX: 729.7MIN: 68.9 / MAX: 734.8MIN: 71.2 / MAX: 731.7MIN: 72.5 / MAX: 731.11. (CXX) g++ options: -O3 -fno-trapping-math -fno-math-errno -march=native -mfma -lgtest

srsRAN Project

Test: PUSCH Processor Benchmark, Throughput Total

OpenBenchmarking.orgMbps, More Is BettersrsRAN Project 23.3Test: PUSCH Processor Benchmark, Throughput Totalcebda15003000450060007500SE +/- 87.81, N = 96547.46774.56898.67079.57122.4MIN: 3614.7 / MAX: 12722MIN: 3650.8 / MAX: 12618.4MIN: 2932.3 / MAX: 13017.6MIN: 4942.3 / MAX: 12824.3MIN: 4599.2 / MAX: 12734.91. (CXX) g++ options: -O3 -fno-trapping-math -fno-math-errno -march=native -mfma -lgtest

srsRAN Project

Test: PUSCH Processor Benchmark, Throughput Thread

OpenBenchmarking.orgMbps, More Is BettersrsRAN Project 23.3Test: PUSCH Processor Benchmark, Throughput Threaddecba714212835SE +/- 0.22, N = 328.828.929.729.829.9MIN: 15.8 / MAX: 52.3MIN: 18.9 / MAX: 52.7MIN: 18.8 / MAX: 52.3MIN: 18.3 / MAX: 53.3MIN: 19.5 / MAX: 52.71. (CXX) g++ options: -O3 -fno-trapping-math -fno-math-errno -march=native -mfma -lgtest

nginx

Connections: 500

OpenBenchmarking.orgRequests Per Second, More Is Betternginx 1.23.2Connections: 500bcdea50K100K150K200K250KSE +/- 1323.62, N = 3246156.11246619.54247581.64248416.85250533.371. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2

Apache HTTP Server

Concurrent Requests: 500

OpenBenchmarking.orgRequests Per Second, More Is BetterApache HTTP Server 2.4.56Concurrent Requests: 500cabde20K40K60K80K100KSE +/- 98.05, N = 377777.0380395.5983834.8184694.7685357.841. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2

oneDNN

Harness: IP Shapes 1D - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUcadeb0.81811.63622.45433.27244.0905SE +/- 0.17695, N = 153.635853.567593.504853.446773.05000MIN: 3.11MIN: 3.02MIN: 2.9MIN: 3.04MIN: 1.61. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: IP Shapes 3D - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUebcad0.6321.2641.8962.5283.16SE +/- 0.03552, N = 32.808692.676992.528482.497572.37479MIN: 2.24MIN: 2.13MIN: 2.05MIN: 2.05MIN: 1.921. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUecadb1.20322.40643.60964.81286.016SE +/- 0.18778, N = 125.347555.307695.193794.875484.62478MIN: 4.19MIN: 3.99MIN: 3.98MIN: 3.78MIN: 2.461. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUcedba0.25950.5190.77851.0381.2975SE +/- 0.002492, N = 31.1533200.9893080.9813610.9784280.872476MIN: 0.92MIN: 0.78MIN: 0.78MIN: 0.77MIN: 0.671. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPUaecbd1.32412.64823.97235.29646.6205SE +/- 0.08402, N = 155.884725.558005.420715.387344.97718MIN: 4.65MIN: 4.37MIN: 4.25MIN: 3.77MIN: 3.921. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPUabecd0.71281.42562.13842.85123.564SE +/- 0.03679, N = 153.167793.046383.021882.913812.83754MIN: 2.49MIN: 2.17MIN: 2.44MIN: 2.28MIN: 2.211. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUadcbe0.0920.1840.2760.3680.46SE +/- 0.000551, N = 30.4087110.4084160.4056770.4029830.400325MIN: 0.36MIN: 0.36MIN: 0.36MIN: 0.36MIN: 0.361. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPUbcdae48121620SE +/- 0.05, N = 314.6314.5414.4914.2714.22MIN: 12.83MIN: 12.86MIN: 12.72MIN: 12.67MIN: 12.71. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPUabced0.1630.3260.4890.6520.815SE +/- 0.002808, N = 30.7244130.7187460.7164190.7122480.711306MIN: 0.66MIN: 0.66MIN: 0.66MIN: 0.66MIN: 0.651. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUcbeda0.09750.1950.29250.390.4875SE +/- 0.019200, N = 150.4335230.3144270.3055030.2961520.239960MIN: 0.18MIN: 0.17MIN: 0.18MIN: 0.18MIN: 0.181. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUaebcd0.09780.19560.29340.39120.489SE +/- 0.003330, N = 150.4346580.4137350.4100290.3974350.391957MIN: 0.33MIN: 0.33MIN: 0.31MIN: 0.32MIN: 0.321. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUadbec0.05150.1030.15450.2060.2575SE +/- 0.001233, N = 30.2289710.2257420.2251970.2193480.219341MIN: 0.2MIN: 0.21MIN: 0.2MIN: 0.21MIN: 0.21. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUacdbe30060090012001500SE +/- 38.79, N = 121216.991209.391182.321155.771120.64MIN: 1149.61MIN: 1153.33MIN: 1123.65MIN: 781.24MIN: 1089.241. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUacebd2004006008001000SE +/- 10.27, N = 15881.23852.58848.65840.96731.10MIN: 840.73MIN: 818.16MIN: 823.74MIN: 756.78MIN: 715.121. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUabdec30060090012001500SE +/- 23.27, N = 141304.571232.871205.381200.191081.70MIN: 1219.16MIN: 1015.69MIN: 1177.67MIN: 1170.19MIN: 10101. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPUabedc0.05150.1030.15450.2060.2575SE +/- 0.002565, N = 30.2287410.2231420.2220200.2194900.217420MIN: 0.19MIN: 0.19MIN: 0.2MIN: 0.2MIN: 0.191. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPUacbed0.10580.21160.31740.42320.529SE +/- 0.003398, N = 110.4703360.4578930.4513930.4462320.440410MIN: 0.36MIN: 0.35MIN: 0.34MIN: 0.35MIN: 0.351. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPUbadce0.10490.20980.31470.41960.5245SE +/- 0.002656, N = 30.4660450.4642690.4625890.4579960.453885MIN: 0.38MIN: 0.38MIN: 0.37MIN: 0.39MIN: 0.41. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUaecdb2004006008001000SE +/- 14.33, N = 15873.15845.73844.36832.57832.34MIN: 841.29MIN: 832.31MIN: 819.84MIN: 807.52MIN: 744.451. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUcabde30060090012001500SE +/- 17.42, N = 151228.771205.291184.141184.121112.04MIN: 1195.53MIN: 1166.28MIN: 1007.85MIN: 1154.2MIN: 1093.031. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

oneDNN

Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.1Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUcdbae2004006008001000SE +/- 9.64, N = 5904.27888.73878.49861.15818.44MIN: 846.06MIN: 874.25MIN: 833.55MIN: 828.35MIN: 804.261. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

Blender

Blend File: BMW27 - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.5Blend File: BMW27 - Compute: CPU-Onlyecbda48121620SE +/- 0.15, N = 414.3014.2114.2014.0414.03

Blender

Blend File: Classroom - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.5Blend File: Classroom - Compute: CPU-Onlycbaed816243240SE +/- 0.30, N = 336.7936.6636.5036.3636.31

Blender

Blend File: Fishy Cat - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.5Blend File: Fishy Cat - Compute: CPU-Onlydcbea510152025SE +/- 0.09, N = 320.1319.9419.7019.5419.36

Blender

Blend File: Barbershop - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.5Blend File: Barbershop - Compute: CPU-Onlyebadc306090120150SE +/- 0.81, N = 3148.11147.73147.25147.18146.59

Blender

Blend File: Pabellon Barcelona - Compute: CPU-Only

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.5Blend File: Pabellon Barcelona - Compute: CPU-Onlyabdce1122334455SE +/- 0.10, N = 348.8147.8447.7347.6547.43


Phoronix Test Suite v10.8.5