xxxs

Intel Core i9-10980XE testing with a ASRock X299 Steel Legend (P1.30 BIOS) and llvmpipe on Ubuntu 22.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2210103-PTS-XXXS677144
Jump To Table - Results

Statistics

Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
A
October 09 2022
  1 Day, 16 Hours, 41 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


xxxsOpenBenchmarking.orgPhoronix Test SuiteIntel Core i9-10980XE @ 4.80GHz (18 Cores / 36 Threads)ASRock X299 Steel Legend (P1.30 BIOS)Intel Sky Lake-E DMI3 Registers32GBSamsung SSD 970 PRO 512GBllvmpipeRealtek ALC1220Intel I219-V + Intel I211Ubuntu 22.045.19.0-051900rc7-generic (x86_64)GNOME Shell 42.2X Server 1.21.1.34.5 Mesa 22.0.1 (LLVM 13.0.1 256 bits)1.2.204GCC 11.2.0ext41024x768ProcessorMotherboardChipsetMemoryDiskGraphicsAudioNetworkOSKernelDesktopDisplay ServerOpenGLVulkanCompilerFile-SystemScreen ResolutionXxxs PerformanceSystem Logs- Transparent Huge Pages: madvise- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: intel_cpufreq schedutil - CPU Microcode: 0x5003302- OpenJDK Runtime Environment (build 11.0.16+8-post-Ubuntu-0ubuntu122.04)- Python 3.10.4- itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Mitigation of Clear buffers; SMT vulnerable + retbleed: Mitigation of Enhanced IBRS + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Mitigation of TSX disabled

xxxssmhasher: wyhashsmhasher: wyhashsmhasher: SHA3-256smhasher: SHA3-256smhasher: Spooky32smhasher: Spooky32smhasher: fasthash32smhasher: fasthash32smhasher: FarmHash128smhasher: FarmHash128smhasher: t1ha2_atoncesmhasher: t1ha2_atoncesmhasher: FarmHash32 x86_64 AVXsmhasher: FarmHash32 x86_64 AVXsmhasher: t1ha0_aes_avx2 x86_64smhasher: t1ha0_aes_avx2 x86_64smhasher: MeowHash x86_64 AES-NIsmhasher: MeowHash x86_64 AES-NIopenfoam: motorBike - Mesh Timeopenfoam: motorBike - Execution Timeopenfoam: drivaerFastback, Small Mesh Size - Mesh Timeopenfoam: drivaerFastback, Small Mesh Size - Execution Timewebp: Defaultwebp: Quality 100webp: Quality 100, Losslesswebp: Quality 100, Highest Compressionwebp2: Defaultwebp2: Quality 75, Compression Effort 7webp2: Quality 95, Compression Effort 7webp2: Quality 100, Compression Effort 5srsran: OFDM_Testsrsran: 4G PHY_DL_Test 100 PRB MIMO 64-QAMsrsran: 4G PHY_DL_Test 100 PRB MIMO 64-QAMsrsran: 4G PHY_DL_Test 100 PRB SISO 64-QAMsrsran: 4G PHY_DL_Test 100 PRB SISO 64-QAMsrsran: 4G PHY_DL_Test 100 PRB MIMO 256-QAMsrsran: 4G PHY_DL_Test 100 PRB MIMO 256-QAMsrsran: 4G PHY_DL_Test 100 PRB SISO 256-QAMsrsran: 4G PHY_DL_Test 100 PRB SISO 256-QAMsrsran: 5G PHY_DL_NR Test 52 PRB SISO 64-QAMsrsran: 5G PHY_DL_NR Test 52 PRB SISO 64-QAMquadray: 1 - 4Kquadray: 2 - 4Kquadray: 3 - 4Kquadray: 5 - 4Kquadray: 1 - 1080pquadray: 2 - 1080pquadray: 3 - 1080pquadray: 5 - 1080paom-av1: Speed 0 Two-Pass - Bosphorus 4Kaom-av1: Speed 4 Two-Pass - Bosphorus 4Kaom-av1: Speed 6 Realtime - Bosphorus 4Kaom-av1: Speed 6 Two-Pass - Bosphorus 4Kaom-av1: Speed 8 Realtime - Bosphorus 4Kaom-av1: Speed 9 Realtime - Bosphorus 4Kaom-av1: Speed 10 Realtime - Bosphorus 4Kaom-av1: Speed 0 Two-Pass - Bosphorus 1080paom-av1: Speed 4 Two-Pass - Bosphorus 1080paom-av1: Speed 6 Realtime - Bosphorus 1080paom-av1: Speed 6 Two-Pass - Bosphorus 1080paom-av1: Speed 8 Realtime - Bosphorus 1080paom-av1: Speed 9 Realtime - Bosphorus 1080paom-av1: Speed 10 Realtime - Bosphorus 1080py-cruncher: 1By-cruncher: 500Monednn: IP Shapes 1D - f32 - CPUonednn: IP Shapes 3D - f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUonednn: IP Shapes 1D - bf16bf16bf16 - CPUonednn: IP Shapes 3D - bf16bf16bf16 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Deconvolution Batch shapes_1d - f32 - CPUonednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - bf16bf16bf16 - CPUonednn: Deconvolution Batch shapes_1d - bf16bf16bf16 - CPUonednn: Deconvolution Batch shapes_3d - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - bf16bf16bf16 - CPUencode-flac: WAV To FLACtensorflow: CPU - 16 - AlexNettensorflow: CPU - 32 - AlexNettensorflow: CPU - 64 - AlexNettensorflow: CPU - 16 - GoogLeNettensorflow: CPU - 16 - ResNet-50tensorflow: CPU - 32 - GoogLeNettensorflow: CPU - 32 - ResNet-50tensorflow: CPU - 64 - GoogLeNettensorflow: CPU - 64 - ResNet-50spacy: en_core_web_lgspacy: en_core_web_trfblender: BMW27 - CPU-Onlyblender: Classroom - CPU-Onlyblender: Fishy Cat - CPU-Onlyblender: Barbershop - CPU-Onlyblender: Pabellon Barcelona - CPU-Onlyrocksdb: Rand Fillrocksdb: Rand Readrocksdb: Update Randrocksdb: Seq Fillrocksdb: Rand Fill Syncrocksdb: Read While Writingrocksdb: Read Rand Write Randhbase: 10000 - Increment - 1hbase: 10000 - Increment - 1hbase: 10000 - Increment - 4hbase: 10000 - Increment - 4hbase: 10000 - Increment - 16hbase: 10000 - Increment - 16hbase: 10000 - Increment - 32hbase: 10000 - Increment - 32hbase: 10000 - Increment - 64hbase: 10000 - Increment - 64hbase: 10000 - Increment - 128hbase: 10000 - Increment - 128hbase: 10000 - Rand Read - 1hbase: 10000 - Rand Read - 1hbase: 10000 - Rand Read - 4hbase: 10000 - Rand Read - 4hbase: 1000000 - Increment - 1hbase: 1000000 - Increment - 1hbase: 1000000 - Increment - 4hbase: 1000000 - Increment - 4hbase: 2000000 - Increment - 1hbase: 2000000 - Increment - 1hbase: 2000000 - Increment - 4hbase: 2000000 - Increment - 4hbase: 10000 - Rand Read - 16hbase: 10000 - Rand Read - 16hbase: 10000 - Rand Read - 32hbase: 10000 - Rand Read - 32hbase: 10000 - Rand Read - 64hbase: 10000 - Rand Read - 64hbase: 10000 - Rand Write - 1hbase: 10000 - Rand Write - 1hbase: 10000 - Rand Write - 4hbase: 10000 - Rand Write - 4hbase: 1000000 - Increment - 16hbase: 1000000 - Increment - 16hbase: 1000000 - Increment - 32hbase: 1000000 - Increment - 32hbase: 1000000 - Increment - 64hbase: 1000000 - Increment - 64hbase: 2000000 - Increment - 16hbase: 2000000 - Increment - 16hbase: 2000000 - Increment - 32hbase: 2000000 - Increment - 32hbase: 2000000 - Increment - 64hbase: 2000000 - Increment - 64hbase: 10000 - Rand Read - 128hbase: 10000 - Rand Read - 128hbase: 10000 - Rand Write - 16hbase: 10000 - Rand Write - 16hbase: 10000 - Rand Write - 32hbase: 10000 - Rand Write - 32hbase: 10000 - Rand Write - 64hbase: 10000 - Rand Write - 64hbase: 1000000 - Rand Read - 1hbase: 1000000 - Rand Read - 1hbase: 1000000 - Rand Read - 4hbase: 1000000 - Rand Read - 4hbase: 2000000 - Increment - 128hbase: 2000000 - Increment - 128hbase: 2000000 - Rand Read - 1hbase: 2000000 - Rand Read - 1hbase: 2000000 - Rand Read - 4hbase: 2000000 - Rand Read - 4hbase: 10000 - Rand Write - 128hbase: 10000 - Rand Write - 128hbase: 1000000 - Rand Read - 16hbase: 1000000 - Rand Read - 16hbase: 1000000 - Rand Read - 32hbase: 1000000 - Rand Read - 32hbase: 1000000 - Rand Read - 64hbase: 1000000 - Rand Read - 64hbase: 1000000 - Rand Write - 1hbase: 1000000 - Rand Write - 1hbase: 1000000 - Rand Write - 4hbase: 1000000 - Rand Write - 4hbase: 2000000 - Rand Read - 16hbase: 2000000 - Rand Read - 16hbase: 2000000 - Rand Read - 32hbase: 2000000 - Rand Read - 32hbase: 2000000 - Rand Read - 64hbase: 2000000 - Rand Read - 64hbase: 2000000 - Rand Write - 1hbase: 2000000 - Rand Write - 1hbase: 2000000 - Rand Write - 4hbase: 2000000 - Rand Write - 4hbase: 10000 - Seq Read - 1hbase: 10000 - Seq Read - 1hbase: 10000 - Seq Read - 4hbase: 10000 - Seq Read - 4hbase: 1000000 - Rand Read - 128hbase: 1000000 - Rand Read - 128hbase: 1000000 - Rand Write - 16hbase: 1000000 - Rand Write - 16hbase: 1000000 - Rand Write - 32hbase: 1000000 - Rand Write - 32hbase: 2000000 - Rand Read - 128hbase: 2000000 - Rand Read - 128hbase: 2000000 - Rand Write - 16hbase: 2000000 - Rand Write - 16hbase: 10000 - Seq Read - 16hbase: 10000 - Seq Read - 16hbase: 10000 - Seq Read - 32hbase: 10000 - Seq Read - 32hbase: 10000 - Seq Read - 64hbase: 10000 - Seq Read - 64hbase: 10000 - Seq Write - 1hbase: 10000 - Seq Write - 1hbase: 10000 - Seq Write - 4hbase: 10000 - Seq Write - 4hbase: 10000 - Async Rand Read - 1hbase: 10000 - Async Rand Read - 1hbase: 10000 - Async Rand Read - 4hbase: 10000 - Async Rand Read - 4hbase: 10000 - Seq Read - 128hbase: 10000 - Seq Read - 128hbase: 10000 - Seq Write - 16hbase: 10000 - Seq Write - 16hbase: 10000 - Seq Write - 32hbase: 10000 - Seq Write - 32hbase: 10000 - Seq Write - 64hbase: 10000 - Seq Write - 64hbase: 1000000 - Seq Read - 4hbase: 1000000 - Seq Read - 4hbase: 2000000 - Seq Read - 1hbase: 2000000 - Seq Read - 1hbase: 2000000 - Seq Read - 4hbase: 2000000 - Seq Read - 4hbase: 10000 - Async Rand Read - 16hbase: 10000 - Async Rand Read - 16hbase: 10000 - Async Rand Read - 32hbase: 10000 - Async Rand Read - 32hbase: 10000 - Async Rand Read - 64hbase: 10000 - Async Rand Read - 64hbase: 10000 - Async Rand Write - 1hbase: 10000 - Async Rand Write - 1hbase: 10000 - Async Rand Write - 4hbase: 10000 - Async Rand Write - 4hbase: 10000 - Seq Write - 128hbase: 10000 - Seq Write - 128hbase: 1000000 - Seq Read - 16hbase: 1000000 - Seq Read - 16hbase: 1000000 - Seq Read - 32hbase: 1000000 - Seq Read - 32hbase: 1000000 - Seq Read - 64hbase: 1000000 - Seq Read - 64hbase: 1000000 - Seq Write - 1hbase: 1000000 - Seq Write - 1hbase: 1000000 - Seq Write - 4hbase: 1000000 - Seq Write - 4hbase: 2000000 - Seq Read - 16hbase: 2000000 - Seq Read - 16hbase: 2000000 - Seq Read - 32hbase: 2000000 - Seq Read - 32hbase: 2000000 - Seq Read - 64hbase: 2000000 - Seq Read - 64hbase: 2000000 - Seq Write - 1hbase: 2000000 - Seq Write - 1hbase: 2000000 - Seq Write - 4hbase: 2000000 - Seq Write - 4hbase: 10000 - Async Rand Read - 128hbase: 10000 - Async Rand Read - 128hbase: 10000 - Async Rand Write - 16hbase: 10000 - Async Rand Write - 16hbase: 10000 - Async Rand Write - 32hbase: 10000 - Async Rand Write - 32hbase: 10000 - Async Rand Write - 64hbase: 10000 - Async Rand Write - 64hbase: 1000000 - Async Rand Read - 1hbase: 1000000 - Async Rand Read - 1hbase: 1000000 - Async Rand Read - 4hbase: 1000000 - Async Rand Read - 4hbase: 1000000 - Seq Read - 128hbase: 1000000 - Seq Read - 128hbase: 1000000 - Seq Write - 16hbase: 1000000 - Seq Write - 16hbase: 1000000 - Seq Write - 32hbase: 1000000 - Seq Write - 32hbase: 1000000 - Seq Write - 64hbase: 1000000 - Seq Write - 64hbase: 2000000 - Async Rand Read - 4hbase: 2000000 - Async Rand Read - 4hbase: 2000000 - Seq Read - 128hbase: 2000000 - Seq Read - 128hbase: 2000000 - Seq Write - 16hbase: 2000000 - Seq Write - 16hbase: 10000 - Async Rand Write - 128hbase: 10000 - Async Rand Write - 128hbase: 1000000 - Async Rand Read - 16hbase: 1000000 - Async Rand Read - 16hbase: 1000000 - Async Rand Read - 64hbase: 1000000 - Async Rand Read - 64hbase: 1000000 - Async Rand Write - 1hbase: 1000000 - Async Rand Write - 1hbase: 1000000 - Async Rand Write - 4hbase: 1000000 - Async Rand Write - 4hbase: 2000000 - Async Rand Read - 16hbase: 2000000 - Async Rand Read - 16hbase: 2000000 - Async Rand Read - 32hbase: 2000000 - Async Rand Read - 32hbase: 2000000 - Async Rand Read - 64hbase: 2000000 - Async Rand Read - 64hbase: 2000000 - Async Rand Write - 1hbase: 2000000 - Async Rand Write - 1hbase: 2000000 - Async Rand Write - 4hbase: 2000000 - Async Rand Write - 4hbase: 2000000 - Async Rand Read - 128hbase: 2000000 - Async Rand Read - 128brl-cad: VGR Performance MetricA29510.6216.954171.022272.34517932.9736.9838747.6925.67519770.2646.8719185.8523.64629486.329.61753384.0923.57647578.5144.63349.2298119.3839.406937264.2368416.8210.511.433.458.260.190.095.25123300000318124.9315.6129.7348137344.1138.5105.566.217.794.584.091.2261.3617.3315.734.810.235.9313.8910.0620.2931.9132.510.6410.2523.4222.1649.9165.4768.1931.30414.7522.544174.877430.5874421.304355.633722.884099.974235.358382.754089.03470.4652250.8069951765.61992.2781788.267.8705813.017811.1024938.6051.339231687.31921.1060.3145942.1754819.314133.42181.27223.34108.4327.41111.5131.25110.6134.3811647243487.54255.35118.6951309.31128503839310277338051308807468536971642615090108590771145483406945359028517652729407819116011661588149252593198311863246127493638580464516792978101137611317153812225764529384521093047789840782897767513673097436442844139145012873796113081411616923118120966528136932701558925536951345636142751733123020840559911664913616881818720035031636945271387812810442015216536919223194127437142268006334193650614185273220816577103834123818883541371039302153017257867263808143771092995581792152656435719624951405127411579210851764938522166213522479124616502241465321417179231596822548630234810906955110169705930661237116537925111721336412381719613704288322119889619151917412736125018294134845452221570753792195135029189532399467764525513824303231126273151885526772480796424539712726354524093614261997496402835125633443824529415542238992672208047969855712619060927046118129384217518835296615603418833052210474OpenBenchmarking.org

SMHasher

SMHasher is a hash function tester supporting various algorithms and able to make use of AVX and other modern CPU instruction set extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: wyhashA6K12K18K24K30K29510.621. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: SHA3-256A4080120160200171.021. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: Spooky32A4K8K12K16K20K17932.971. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: fasthash32A2K4K6K8K10K8747.691. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: FarmHash128A4K8K12K16K20K19770.261. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: t1ha2_atonceA4K8K12K16K20K19185.851. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: FarmHash32 x86_64 AVXA6K12K18K24K30K29486.31. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: t1ha0_aes_avx2 x86_64A11K22K33K44K55K53384.091. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects

OpenBenchmarking.orgMiB/sec, More Is BetterSMHasher 2022-08-22Hash: MeowHash x86_64 AES-NIA10K20K30K40K50K47578.511. (CXX) g++ options: -march=native -O3 -flto -fno-fat-lto-objects

OpenFOAM

OpenFOAM is the leading free, open-source software for computational fluid dynamics (CFD). This test profile currently uses the drivaerFastback test case for analyzing automotive aerodynamics or alternatively the older motorBike input. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenFOAM 10Input: motorBike - Mesh TimeA112233445549.231. (CXX) g++ options: -std=c++14 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenFOAM 10Input: motorBike - Execution TimeA306090120150119.381. (CXX) g++ options: -std=c++14 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenFOAM 10Input: drivaerFastback, Small Mesh Size - Mesh TimeA91827364539.411. (CXX) g++ options: -std=c++14 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm

OpenBenchmarking.orgSeconds, Fewer Is BetterOpenFOAM 10Input: drivaerFastback, Small Mesh Size - Execution TimeA60120180240300264.241. (CXX) g++ options: -std=c++14 -m64 -O3 -ftemplate-depth-100 -fPIC -fuse-ld=bfd -Xlinker --add-needed --no-as-needed -lfoamToVTK -ldynamicMesh -llagrangian -lgenericPatchFields -lfileFormats -lOpenFOAM -ldl -lm

WebP Image Encode

This is a test of Google's libwebp with the cwebp image encode utility and using a sample 6000x4000 pixel JPEG image as the input. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterWebP Image Encode 1.2.4Encode Settings: DefaultA4812162016.821. (CC) gcc options: -fvisibility=hidden -O2 -lm

OpenBenchmarking.orgMP/s, More Is BetterWebP Image Encode 1.2.4Encode Settings: Quality 100A369121510.511. (CC) gcc options: -fvisibility=hidden -O2 -lm

OpenBenchmarking.orgMP/s, More Is BetterWebP Image Encode 1.2.4Encode Settings: Quality 100, LosslessA0.32180.64360.96541.28721.6091.431. (CC) gcc options: -fvisibility=hidden -O2 -lm

OpenBenchmarking.orgMP/s, More Is BetterWebP Image Encode 1.2.4Encode Settings: Quality 100, Highest CompressionA0.77631.55262.32893.10523.88153.451. (CC) gcc options: -fvisibility=hidden -O2 -lm

WebP2 Image Encode

This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterWebP2 Image Encode 20220823Encode Settings: DefaultA2468108.261. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl

OpenBenchmarking.orgMP/s, More Is BetterWebP2 Image Encode 20220823Encode Settings: Quality 75, Compression Effort 7A0.04280.08560.12840.17120.2140.191. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl

OpenBenchmarking.orgMP/s, More Is BetterWebP2 Image Encode 20220823Encode Settings: Quality 95, Compression Effort 7A0.02030.04060.06090.08120.10150.091. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl

OpenBenchmarking.orgMP/s, More Is BetterWebP2 Image Encode 20220823Encode Settings: Quality 100, Compression Effort 5A1.18132.36263.54394.72525.90655.251. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSamples / Second, More Is BettersrsRAN 22.04.1Test: OFDM_TestA30M60M90M120M150M1233000001. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -mavx512f -mavx512cd -mavx512bw -mavx512dq -ldl -lpthread -lm

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB MIMO 64-QAMA701402102803503181. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -mavx512f -mavx512cd -mavx512bw -mavx512dq -ldl -lpthread -lm

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB MIMO 64-QAMA306090120150124.91. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -mavx512f -mavx512cd -mavx512bw -mavx512dq -ldl -lpthread -lm

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB SISO 64-QAMA70140210280350315.61. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -mavx512f -mavx512cd -mavx512bw -mavx512dq -ldl -lpthread -lm

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB SISO 64-QAMA306090120150129.71. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -mavx512f -mavx512cd -mavx512bw -mavx512dq -ldl -lpthread -lm

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB MIMO 256-QAMA801602403204003481. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -mavx512f -mavx512cd -mavx512bw -mavx512dq -ldl -lpthread -lm

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB MIMO 256-QAMA3060901201501371. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -mavx512f -mavx512cd -mavx512bw -mavx512dq -ldl -lpthread -lm

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB SISO 256-QAMA70140210280350344.11. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -mavx512f -mavx512cd -mavx512bw -mavx512dq -ldl -lpthread -lm

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB SISO 256-QAMA306090120150138.51. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -mavx512f -mavx512cd -mavx512bw -mavx512dq -ldl -lpthread -lm

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 22.04.1Test: 5G PHY_DL_NR Test 52 PRB SISO 64-QAMA20406080100105.51. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -mavx512f -mavx512cd -mavx512bw -mavx512dq -ldl -lpthread -lm

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 22.04.1Test: 5G PHY_DL_NR Test 52 PRB SISO 64-QAMA153045607566.21. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -mavx512f -mavx512cd -mavx512bw -mavx512dq -ldl -lpthread -lm

QuadRay

VectorChief's QuadRay is a real-time ray-tracing engine written to support SIMD across ARM, MIPS, PPC, and x86/x86_64 processors. QuadRay supports SSE/SSE2/SSE4 and AVX/AVX2/AVX-512 usage on Intel/AMD CPUs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 1 - Resolution: 4KA4812162017.791. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 2 - Resolution: 4KA1.03052.0613.09154.1225.15254.581. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 3 - Resolution: 4KA0.92031.84062.76093.68124.60154.091. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 5 - Resolution: 4KA0.27450.5490.82351.0981.37251.221. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 1 - Resolution: 1080pA142842567061.361. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 2 - Resolution: 1080pA4812162017.331. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 3 - Resolution: 1080pA4812162015.731. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

OpenBenchmarking.orgFPS, More Is BetterQuadRay 2022.05.25Scene: 5 - Resolution: 1080pA1.08232.16463.24694.32925.41154.811. (CXX) g++ options: -O3 -pthread -lm -lstdc++ -lX11 -lXext -lpthread

AOM AV1

This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google as the AV1 Codec Library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.5Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 4KA0.05180.10360.15540.20720.2590.231. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.5Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 4KA1.33432.66864.00295.33726.67155.931. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.5Encoder Mode: Speed 6 Realtime - Input: Bosphorus 4KA4812162013.891. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.5Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 4KA369121510.061. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.5Encoder Mode: Speed 8 Realtime - Input: Bosphorus 4KA51015202520.291. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.5Encoder Mode: Speed 9 Realtime - Input: Bosphorus 4KA71421283531.911. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.5Encoder Mode: Speed 10 Realtime - Input: Bosphorus 4KA81624324032.511. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.5Encoder Mode: Speed 0 Two-Pass - Input: Bosphorus 1080pA0.1440.2880.4320.5760.720.641. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.5Encoder Mode: Speed 4 Two-Pass - Input: Bosphorus 1080pA369121510.251. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.5Encoder Mode: Speed 6 Realtime - Input: Bosphorus 1080pA61218243023.421. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.5Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 1080pA51015202522.161. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.5Encoder Mode: Speed 8 Realtime - Input: Bosphorus 1080pA112233445549.911. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.5Encoder Mode: Speed 9 Realtime - Input: Bosphorus 1080pA153045607565.471. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

OpenBenchmarking.orgFrames Per Second, More Is BetterAOM AV1 3.5Encoder Mode: Speed 10 Realtime - Input: Bosphorus 1080pA153045607568.191. (CXX) g++ options: -O3 -std=c++11 -U_FORTIFY_SOURCE -lm

Y-Cruncher

Y-Cruncher is a multi-threaded Pi benchmark capable of computing Pi to trillions of digits. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterY-Cruncher 0.7.10.9513Pi Digits To Calculate: 1BA71421283531.30

OpenBenchmarking.orgSeconds, Fewer Is BetterY-Cruncher 0.7.10.9513Pi Digits To Calculate: 500MA4812162014.75

oneDNN

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUA0.57241.14481.71722.28962.8622.54417MIN: 2.381. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUA1.09742.19483.29224.38965.4874.87743MIN: 4.781. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUA0.13220.26440.39660.52880.6610.587442MIN: 0.531. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUA0.29350.5870.88051.1741.46751.30435MIN: 1.221. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPUA1.26762.53523.80285.07046.3385.63372MIN: 5.471. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPUA0.64891.29781.94672.59563.24452.88409MIN: 2.741. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUA36912159.97423MIN: 9.891. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPUA1.20562.41123.61684.82246.0285.35838MIN: 4.041. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPUA0.61971.23941.85912.47883.09852.75408MIN: 2.721. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUA36912159.0347MIN: 8.971. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUA0.10470.20940.31410.41880.52350.465225MIN: 0.451. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUA0.18160.36320.54480.72640.9080.806995MIN: 0.771. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUA4008001200160020001765.61MIN: 1738.61. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUA2004006008001000992.28MIN: 958.521. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUA4008001200160020001788.26MIN: 1750.951. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPUA2468107.87058MIN: 7.611. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPUA369121513.02MIN: 12.791. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPUA369121511.10MIN: 10.951. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUA2004006008001000938.61MIN: 913.541. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPUA0.30130.60260.90391.20521.50651.33923MIN: 1.31. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUA4008001200160020001687.31MIN: 1677.681. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUA2004006008001000921.11MIN: 910.951. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPUA0.07080.14160.21240.28320.3540.314594MIN: 0.31. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.7Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPUA0.48950.9791.46851.9582.44752.17548MIN: 1.951. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl -lpthread

FLAC Audio Encoding

OpenBenchmarking.orgSeconds, Fewer Is BetterFLAC Audio Encoding 1.4WAV To FLACA51015202519.311. (CXX) g++ options: -O3 -fvisibility=hidden -logg -lm

TensorFlow

This is a benchmark of the TensorFlow deep learning framework using the TensorFlow reference benchmarks (tensorflow/benchmarks with tf_cnn_benchmarks.py). Note with the Phoronix Test Suite there is also pts/tensorflow-lite for benchmarking the TensorFlow Lite binaries too. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 16 - Model: AlexNetA306090120150133.42

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 32 - Model: AlexNetA4080120160200181.27

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 64 - Model: AlexNetA50100150200250223.34

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 16 - Model: GoogLeNetA20406080100108.43

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 16 - Model: ResNet-50A61218243027.41

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 32 - Model: GoogLeNetA20406080100111.51

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 32 - Model: ResNet-50A71421283531.25

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 64 - Model: GoogLeNetA20406080100110.61

OpenBenchmarking.orgimages/sec, More Is BetterTensorFlow 2.10Device: CPU - Batch Size: 64 - Model: ResNet-50A81624324034.38

spaCy

The spaCy library is an open-source solution for advanced neural language processing (NLP). The spaCy library leverages Python and is a leading neural language processing solution. This test profile times the spaCy CPU performance with various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgtokens/sec, More Is BetterspaCy 3.4.1Model: en_core_web_lgA2K4K6K8K10K11647

OpenBenchmarking.orgtokens/sec, More Is BetterspaCy 3.4.1Model: en_core_web_trfA50010001500200025002434

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.3Blend File: BMW27 - Compute: CPU-OnlyA2040608010087.54

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.3Blend File: Classroom - Compute: CPU-OnlyA60120180240300255.35

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.3Blend File: Fishy Cat - Compute: CPU-OnlyA306090120150118.6

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.3Blend File: Barbershop - Compute: CPU-OnlyA2004006008001000951

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.3Blend File: Pabellon Barcelona - Compute: CPU-OnlyA70140210280350309.3

Facebook RocksDB

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.5.3Test: Random FillA200K400K600K800K1000K11285031. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.5.3Test: Random ReadA20M40M60M80M100M839310271. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.5.3Test: Update RandomA160K320K480K640K800K7338051. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.5.3Test: Sequential FillA300K600K900K1200K1500K13088071. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.5.3Test: Random Fill SyncA1000200030004000500046851. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.5.3Test: Read While WritingA800K1600K2400K3200K4000K36971641. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.5.3Test: Read Random Write RandomA600K1200K1800K2400K3000K26150901. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

Apache HBase

This is a benchmark of the Apache HBase non-relational distributed database system inspired from Google's Bigtable. Learn more via the OpenBenchmarking.org test page.

Rows: 10000 - Test: Scan - Clients: 1

A: The test run did not produce a result.

Rows: 10000 - Test: Scan - Clients: 4

A: The test run did not produce a result.

Rows: 10000 - Test: Scan - Clients: 16

A: The test run did not produce a result.

Rows: 10000 - Test: Scan - Clients: 32

A: The test run did not produce a result.

Rows: 10000 - Test: Scan - Clients: 64

A: The test run did not produce a result.

Rows: 10000 - Test: Scan - Clients: 128

A: The test run did not produce a result.

Rows: 10000 - Test: Scan - Clients: 256

A: The test run did not produce a result.

Rows: 10000 - Test: Scan - Clients: 500

A: The test run did not produce a result.

Rows: 1000000 - Test: Scan - Clients: 1

A: The test run did not produce a result.

Rows: 1000000 - Test: Scan - Clients: 4

A: The test run did not produce a result.

Rows: 2000000 - Test: Scan - Clients: 1

A: The test run did not produce a result.

Rows: 2000000 - Test: Scan - Clients: 4

A: The test run did not produce a result.

Rows: 1000000 - Test: Scan - Clients: 16

A: The test run did not produce a result.

Rows: 1000000 - Test: Scan - Clients: 32

A: The test run did not produce a result.

Rows: 1000000 - Test: Scan - Clients: 64

A: The test run did not produce a result.

Rows: 2000000 - Test: Scan - Clients: 16

A: The test run did not produce a result.

Rows: 2000000 - Test: Scan - Clients: 32

A: The test run did not produce a result.

Rows: 2000000 - Test: Scan - Clients: 64

A: The test run did not produce a result.

Rows: 1000000 - Test: Scan - Clients: 128

A: The test run did not produce a result.

Rows: 1000000 - Test: Scan - Clients: 256

A: The test run did not produce a result.

Rows: 1000000 - Test: Scan - Clients: 500

A: The test run did not produce a result.

Rows: 2000000 - Test: Scan - Clients: 128

A: The test run did not produce a result.

Rows: 2000000 - Test: Scan - Clients: 256

A: The test run did not produce a result.

Rows: 2000000 - Test: Scan - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Increment - Clients: 1A20040060080010001085

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Increment - Clients: 1A2004006008001000907

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Increment - Clients: 4A150030004500600075007114

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Increment - Clients: 4A120240360480600548

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Increment - Clients: 16A7K14K21K28K35K34069

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Increment - Clients: 16A100200300400500453

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Increment - Clients: 32A13K26K39K52K65K59028

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Increment - Clients: 32A110220330440550517

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Increment - Clients: 64A14K28K42K56K70K65272

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Increment - Clients: 64A2004006008001000940

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Increment - Clients: 128A20K40K60K80K100K78191

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Increment - Clients: 128A300600900120015001601

Rows: 10000 - Test: Increment - Clients: 256

A: The test run did not produce a result.

Rows: 10000 - Test: Increment - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Read - Clients: 1A4008001200160020001661

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Read - Clients: 1A130260390520650588

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Read - Clients: 4A3K6K9K12K15K14925

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Read - Clients: 4A60120180240300259

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Increment - Clients: 1A70014002100280035003198

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Increment - Clients: 1A70140210280350311

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Increment - Clients: 4A2K4K6K8K10K8632

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Increment - Clients: 4A100200300400500461

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Increment - Clients: 1A60012001800240030002749

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Increment - Clients: 1A80160240320400363

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Increment - Clients: 4A2K4K6K8K10K8580

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Increment - Clients: 4A100200300400500464

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Read - Clients: 16A11K22K33K44K55K51679

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Read - Clients: 16A60120180240300297

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Read - Clients: 32A20K40K60K80K100K81011

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Read - Clients: 32A80160240320400376

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Read - Clients: 64A20K40K60K80K100K113171

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Read - Clients: 64A120240360480600538

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Write - Clients: 1A3K6K9K12K15K12225

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Write - Clients: 1A2040608010076

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Write - Clients: 4A10K20K30K40K50K45293

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Write - Clients: 4A2040608010084

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Increment - Clients: 16A11K22K33K44K55K52109

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Increment - Clients: 16A70140210280350304

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Increment - Clients: 32A20K40K60K80K100K77898

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Increment - Clients: 32A90180270360450407

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Increment - Clients: 64A20K40K60K80K100K82897

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Increment - Clients: 64A170340510680850767

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Increment - Clients: 16A11K22K33K44K55K51367

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Increment - Clients: 16A70140210280350309

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Increment - Clients: 32A16K32K48K64K80K74364

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Increment - Clients: 32A90180270360450428

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Increment - Clients: 64A9K18K27K36K45K44139

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Increment - Clients: 64A300600900120015001450

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Read - Clients: 128A30K60K90K120K150K128737

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Read - Clients: 128A2004006008001000961

Rows: 10000 - Test: Random Read - Clients: 256

A: The test run did not produce a result.

Rows: 10000 - Test: Random Read - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Write - Clients: 16A30K60K90K120K150K130814

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Write - Clients: 16A306090120150116

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Write - Clients: 32A40K80K120K160K200K169231

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Write - Clients: 32A4080120160200181

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Write - Clients: 64A40K80K120K160K200K209665

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Write - Clients: 64A60120180240300281

Rows: 1000000 - Test: Increment - Clients: 128

A: The test run did not produce a result.

Rows: 1000000 - Test: Increment - Clients: 256

A: The test run did not produce a result.

Rows: 1000000 - Test: Increment - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Read - Clients: 1A80016002400320040003693

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Read - Clients: 1A60120180240300270

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Read - Clients: 4A3K6K9K12K15K15589

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Read - Clients: 4A60120180240300255

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Increment - Clients: 128A8K16K24K32K40K36951

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Increment - Clients: 128A70014002100280035003456

Rows: 2000000 - Test: Increment - Clients: 256

A: The test run did not produce a result.

Rows: 2000000 - Test: Increment - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Read - Clients: 1A80016002400320040003614

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Read - Clients: 1A60120180240300275

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Read - Clients: 4A4K8K12K16K20K17331

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Read - Clients: 4A50100150200250230

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Write - Clients: 128A40K80K120K160K200K208405

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Random Write - Clients: 128A130260390520650599

Rows: 10000 - Test: Random Write - Clients: 256

A: The test run did not produce a result.

Rows: 10000 - Test: Random Write - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Read - Clients: 16A20K40K60K80K100K116649

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Read - Clients: 16A306090120150136

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Read - Clients: 32A40K80K120K160K200K168818

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Read - Clients: 32A4080120160200187

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Read - Clients: 64A40K80K120K160K200K200350

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Read - Clients: 64A70140210280350316

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Write - Clients: 1A8K16K24K32K40K36945

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Write - Clients: 1A61218243027

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Write - Clients: 4A30K60K90K120K150K138781

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Write - Clients: 4A71421283528

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Read - Clients: 16A20K40K60K80K100K104420

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Read - Clients: 16A306090120150152

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Read - Clients: 32A40K80K120K160K200K165369

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Read - Clients: 32A4080120160200192

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Read - Clients: 64A50K100K150K200K250K231941

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Read - Clients: 64A60120180240300274

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Write - Clients: 1A8K16K24K32K40K37142

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Write - Clients: 1A61218243026

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Write - Clients: 4A20K40K60K80K100K80063

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Write - Clients: 4A81624324034

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Read - Clients: 1A4008001200160020001936

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Read - Clients: 1A110220330440550506

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Read - Clients: 4A3K6K9K12K15K14185

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Read - Clients: 4A60120180240300273

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Read - Clients: 128A50K100K150K200K250K220816

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Read - Clients: 128A120240360480600577

Rows: 1000000 - Test: Random Read - Clients: 256

A: The test run did not produce a result.

Rows: 1000000 - Test: Random Read - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Write - Clients: 16A20K40K60K80K100K103834

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Write - Clients: 16A306090120150123

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Write - Clients: 32A20K40K60K80K100K81888

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Random Write - Clients: 32A80160240320400354

Rows: 1000000 - Test: Random Write - Clients: 64

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Read - Clients: 128A30K60K90K120K150K137103

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Read - Clients: 128A2004006008001000930

Rows: 2000000 - Test: Random Read - Clients: 256

A: The test run did not produce a result.

Rows: 2000000 - Test: Random Read - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Write - Clients: 16A50K100K150K200K250K215301

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Random Write - Clients: 16A163248648072

Rows: 2000000 - Test: Random Write - Clients: 32

A: The test run did not produce a result.

Rows: 2000000 - Test: Random Write - Clients: 64

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Read - Clients: 16A12K24K36K48K60K57867

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Read - Clients: 16A60120180240300263

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Read - Clients: 32A20K40K60K80K100K80814

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Read - Clients: 32A80160240320400377

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Read - Clients: 64A20K40K60K80K100K109299

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Read - Clients: 64A120240360480600558

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Write - Clients: 1A4K8K12K16K20K17921

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Write - Clients: 1A122436486052

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Write - Clients: 4A14K28K42K56K70K65643

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Write - Clients: 4A132639526557

Rows: 1000000 - Test: Random Write - Clients: 128

A: The test run did not produce a result.

Rows: 1000000 - Test: Random Write - Clients: 256

A: The test run did not produce a result.

Rows: 1000000 - Test: Random Write - Clients: 500

A: The test run did not produce a result.

Rows: 2000000 - Test: Random Write - Clients: 128

A: The test run did not produce a result.

Rows: 2000000 - Test: Random Write - Clients: 256

A: The test run did not produce a result.

Rows: 2000000 - Test: Random Write - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Read - Clients: 1A4008001200160020001962

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Read - Clients: 1A110220330440550495

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Read - Clients: 4A3K6K9K12K15K14051

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Read - Clients: 4A60120180240300274

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Read - Clients: 128A20K40K60K80K100K115792

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Read - Clients: 128A20040060080010001085

Rows: 10000 - Test: Sequential Read - Clients: 256

A: The test run did not produce a result.

Rows: 10000 - Test: Sequential Read - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Write - Clients: 16A40K80K120K160K200K176493

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Write - Clients: 16A2040608010085

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Write - Clients: 32A50K100K150K200K250K221662

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Write - Clients: 32A306090120150135

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Write - Clients: 64A50K100K150K200K250K224791

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Write - Clients: 64A50100150200250246

Rows: 1000000 - Test: Sequential Read - Clients: 1

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Read - Clients: 4A4K8K12K16K20K16502

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Read - Clients: 4A50100150200250241

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Read - Clients: 1A100020003000400050004653

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Read - Clients: 1A50100150200250214

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Read - Clients: 4A4K8K12K16K20K17179

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Read - Clients: 4A50100150200250231

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Read - Clients: 16A13K26K39K52K65K59682

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Read - Clients: 16A60120180240300254

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Read - Clients: 32A20K40K60K80K100K86302

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Read - Clients: 32A80160240320400348

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Read - Clients: 64A20K40K60K80K100K109069

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Read - Clients: 64A120240360480600551

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Write - Clients: 1A20040060080010001016

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Write - Clients: 1A2004006008001000970

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Write - Clients: 4A130026003900520065005930

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Write - Clients: 4A140280420560700661

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Write - Clients: 128A50K100K150K200K250K237116

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Sequential Write - Clients: 128A120240360480600537

Rows: 10000 - Test: Sequential Write - Clients: 256

A: The test run did not produce a result.

Rows: 10000 - Test: Sequential Write - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Read - Clients: 16A20K40K60K80K100K92511

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Read - Clients: 16A4080120160200172

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Read - Clients: 32A30K60K90K120K150K133641

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Read - Clients: 32A50100150200250238

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Read - Clients: 64A40K80K120K160K200K171961

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Read - Clients: 64A80160240320400370

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Write - Clients: 1A9K18K27K36K45K42883

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Write - Clients: 1A51015202522

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Write - Clients: 4A30K60K90K120K150K119889

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Write - Clients: 4A142842567061

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Read - Clients: 16A20K40K60K80K100K91519

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Read - Clients: 16A4080120160200174

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Read - Clients: 32A30K60K90K120K150K127361

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Read - Clients: 32A50100150200250250

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Read - Clients: 64A40K80K120K160K200K182941

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Read - Clients: 64A80160240320400348

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Write - Clients: 1A10K20K30K40K50K45452

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Write - Clients: 1A51015202522

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Write - Clients: 4A30K60K90K120K150K157075

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Write - Clients: 4A91827364537

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Read - Clients: 128A20K40K60K80K100K92195

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Read - Clients: 128A300600900120015001350

Rows: 10000 - Test: Async Random Read - Clients: 256

A: The test run did not produce a result.

Rows: 10000 - Test: Async Random Read - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Write - Clients: 16A6K12K18K24K30K29189

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Write - Clients: 16A120240360480600532

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Write - Clients: 32A9K18K27K36K45K39946

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Write - Clients: 32A2004006008001000776

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Write - Clients: 64A10K20K30K40K50K45255

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Write - Clients: 64A300600900120015001382

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Async Random Read - Clients: 1A90018002700360045004303

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Async Random Read - Clients: 1A50100150200250231

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Async Random Read - Clients: 4A3K6K9K12K15K12627

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Async Random Read - Clients: 4A70140210280350315

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Read - Clients: 128A40K80K120K160K200K188552

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Read - Clients: 128A150300450600750677

Rows: 1000000 - Test: Sequential Read - Clients: 256

A: The test run did not produce a result.

Rows: 1000000 - Test: Sequential Read - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Write - Clients: 16A50K100K150K200K250K248079

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Write - Clients: 16A142842567064

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Write - Clients: 32A50K100K150K200K250K245397

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Write - Clients: 32A306090120150127

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Write - Clients: 64A60K120K180K240K300K263545

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Sequential Write - Clients: 64A50100150200250240

Rows: 2000000 - Test: Async Random Read - Clients: 1

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Async Random Read - Clients: 4A2K4K6K8K10K9361

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Async Random Read - Clients: 4A90180270360450426

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Read - Clients: 128A40K80K120K160K200K199749

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Read - Clients: 128A140280420560700640

Rows: 2000000 - Test: Sequential Read - Clients: 256

A: The test run did not produce a result.

Rows: 2000000 - Test: Sequential Read - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Write - Clients: 16A60K120K180K240K300K283512

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Sequential Write - Clients: 16A132639526556

Rows: 2000000 - Test: Sequential Write - Clients: 32

A: The test run did not produce a result.

Rows: 2000000 - Test: Sequential Write - Clients: 64

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Write - Clients: 128A70014002100280035003344

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 10000 - Test: Async Random Write - Clients: 128A8K16K24K32K40K38245

Rows: 10000 - Test: Async Random Write - Clients: 256

A: The test run did not produce a result.

Rows: 10000 - Test: Async Random Write - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Async Random Read - Clients: 16A6K12K18K24K30K29415

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Async Random Read - Clients: 16A120240360480600542

Rows: 1000000 - Test: Async Random Read - Clients: 32

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Async Random Read - Clients: 64A5K10K15K20K25K23899

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Async Random Read - Clients: 64A60012001800240030002672

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Async Random Write - Clients: 1A4008001200160020002080

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Async Random Write - Clients: 1A100200300400500479

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Async Random Write - Clients: 4A150030004500600075006985

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 1000000 - Test: Async Random Write - Clients: 4A120240360480600571

Rows: 1000000 - Test: Sequential Write - Clients: 128

A: The test run did not produce a result.

Rows: 1000000 - Test: Sequential Write - Clients: 256

A: The test run did not produce a result.

Rows: 1000000 - Test: Sequential Write - Clients: 500

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Async Random Read - Clients: 16A6K12K18K24K30K26190

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Async Random Read - Clients: 16A130260390520650609

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Async Random Read - Clients: 32A6K12K18K24K30K27046

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Async Random Read - Clients: 32A300600900120015001181

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Async Random Read - Clients: 64A6K12K18K24K30K29384

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Async Random Read - Clients: 64A50010001500200025002175

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Async Random Write - Clients: 1A4008001200160020001883

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Async Random Write - Clients: 1A110220330440550529

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Async Random Write - Clients: 4A140028004200560070006615

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Async Random Write - Clients: 4A130260390520650603

Rows: 2000000 - Test: Sequential Write - Clients: 128

A: The test run did not produce a result.

Rows: 2000000 - Test: Sequential Write - Clients: 256

A: The test run did not produce a result.

Rows: 2000000 - Test: Sequential Write - Clients: 500

A: The test run did not produce a result.

Rows: 1000000 - Test: Async Random Read - Clients: 128

A: The test run did not produce a result.

Rows: 1000000 - Test: Async Random Read - Clients: 256

A: The test run did not produce a result.

Rows: 1000000 - Test: Async Random Read - Clients: 500

A: The test run did not produce a result.

Rows: 1000000 - Test: Async Random Write - Clients: 16

A: The test run did not produce a result.

Rows: 1000000 - Test: Async Random Write - Clients: 32

A: The test run did not produce a result.

Rows: 1000000 - Test: Async Random Write - Clients: 64

A: The test run did not produce a result.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Async Random Read - Clients: 128A9K18K27K36K45K41883

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.5.0Rows: 2000000 - Test: Async Random Read - Clients: 128A70014002100280035003052

Rows: 2000000 - Test: Async Random Read - Clients: 256

A: The test run did not produce a result.

Rows: 2000000 - Test: Async Random Read - Clients: 500

A: The test run did not produce a result.

Rows: 2000000 - Test: Async Random Write - Clients: 16

A: The test run did not produce a result.

Rows: 2000000 - Test: Async Random Write - Clients: 32

A: The test run did not produce a result.

Rows: 2000000 - Test: Async Random Write - Clients: 64

A: The test run did not produce a result. E: cat: output.log: No such file or directory

BRL-CAD

BRL-CAD is a cross-platform, open-source solid modeling system with built-in benchmark mode. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgVGR Performance Metric, More Is BetterBRL-CAD 7.32.6VGR Performance MetricA50K100K150K200K250K2104741. (CXX) g++ options: -std=c++11 -pipe -fvisibility=hidden -fno-strict-aliasing -fno-common -fexceptions -ftemplate-depth-128 -m64 -ggdb3 -O3 -fipa-pta -fstrength-reduce -finline-functions -flto -pedantic -ldl -lm

313 Results Shown

SMHasher:
  wyhash
  SHA3-256
  Spooky32
  fasthash32
  FarmHash128
  t1ha2_atonce
  FarmHash32 x86_64 AVX
  t1ha0_aes_avx2 x86_64
  MeowHash x86_64 AES-NI
OpenFOAM:
  motorBike - Mesh Time
  motorBike - Execution Time
  drivaerFastback, Small Mesh Size - Mesh Time
  drivaerFastback, Small Mesh Size - Execution Time
WebP Image Encode:
  Default
  Quality 100
  Quality 100, Lossless
  Quality 100, Highest Compression
WebP2 Image Encode:
  Default
  Quality 75, Compression Effort 7
  Quality 95, Compression Effort 7
  Quality 100, Compression Effort 5
srsRAN:
  OFDM_Test
  4G PHY_DL_Test 100 PRB MIMO 64-QAM
  4G PHY_DL_Test 100 PRB MIMO 64-QAM
  4G PHY_DL_Test 100 PRB SISO 64-QAM
  4G PHY_DL_Test 100 PRB SISO 64-QAM
  4G PHY_DL_Test 100 PRB MIMO 256-QAM
  4G PHY_DL_Test 100 PRB MIMO 256-QAM
  4G PHY_DL_Test 100 PRB SISO 256-QAM
  4G PHY_DL_Test 100 PRB SISO 256-QAM
  5G PHY_DL_NR Test 52 PRB SISO 64-QAM
  5G PHY_DL_NR Test 52 PRB SISO 64-QAM
QuadRay:
  1 - 4K
  2 - 4K
  3 - 4K
  5 - 4K
  1 - 1080p
  2 - 1080p
  3 - 1080p
  5 - 1080p
AOM AV1:
  Speed 0 Two-Pass - Bosphorus 4K
  Speed 4 Two-Pass - Bosphorus 4K
  Speed 6 Realtime - Bosphorus 4K
  Speed 6 Two-Pass - Bosphorus 4K
  Speed 8 Realtime - Bosphorus 4K
  Speed 9 Realtime - Bosphorus 4K
  Speed 10 Realtime - Bosphorus 4K
  Speed 0 Two-Pass - Bosphorus 1080p
  Speed 4 Two-Pass - Bosphorus 1080p
  Speed 6 Realtime - Bosphorus 1080p
  Speed 6 Two-Pass - Bosphorus 1080p
  Speed 8 Realtime - Bosphorus 1080p
  Speed 9 Realtime - Bosphorus 1080p
  Speed 10 Realtime - Bosphorus 1080p
Y-Cruncher:
  1B
  500M
oneDNN:
  IP Shapes 1D - f32 - CPU
  IP Shapes 3D - f32 - CPU
  IP Shapes 1D - u8s8f32 - CPU
  IP Shapes 3D - u8s8f32 - CPU
  IP Shapes 1D - bf16bf16bf16 - CPU
  IP Shapes 3D - bf16bf16bf16 - CPU
  Convolution Batch Shapes Auto - f32 - CPU
  Deconvolution Batch shapes_1d - f32 - CPU
  Deconvolution Batch shapes_3d - f32 - CPU
  Convolution Batch Shapes Auto - u8s8f32 - CPU
  Deconvolution Batch shapes_1d - u8s8f32 - CPU
  Deconvolution Batch shapes_3d - u8s8f32 - CPU
  Recurrent Neural Network Training - f32 - CPU
  Recurrent Neural Network Inference - f32 - CPU
  Recurrent Neural Network Training - u8s8f32 - CPU
  Convolution Batch Shapes Auto - bf16bf16bf16 - CPU
  Deconvolution Batch shapes_1d - bf16bf16bf16 - CPU
  Deconvolution Batch shapes_3d - bf16bf16bf16 - CPU
  Recurrent Neural Network Inference - u8s8f32 - CPU
  Matrix Multiply Batch Shapes Transformer - f32 - CPU
  Recurrent Neural Network Training - bf16bf16bf16 - CPU
  Recurrent Neural Network Inference - bf16bf16bf16 - CPU
  Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPU
  Matrix Multiply Batch Shapes Transformer - bf16bf16bf16 - CPU
FLAC Audio Encoding
TensorFlow:
  CPU - 16 - AlexNet
  CPU - 32 - AlexNet
  CPU - 64 - AlexNet
  CPU - 16 - GoogLeNet
  CPU - 16 - ResNet-50
  CPU - 32 - GoogLeNet
  CPU - 32 - ResNet-50
  CPU - 64 - GoogLeNet
  CPU - 64 - ResNet-50
spaCy:
  en_core_web_lg
  en_core_web_trf
Blender:
  BMW27 - CPU-Only
  Classroom - CPU-Only
  Fishy Cat - CPU-Only
  Barbershop - CPU-Only
  Pabellon Barcelona - CPU-Only
Facebook RocksDB:
  Rand Fill
  Rand Read
  Update Rand
  Seq Fill
  Rand Fill Sync
  Read While Writing
  Read Rand Write Rand
Apache HBase:
  10000 - Increment - 1:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Increment - 4:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Increment - 16:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Increment - 32:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Increment - 64:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Increment - 128:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Rand Read - 1:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Rand Read - 4:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Increment - 1:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Increment - 4:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Increment - 1:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Increment - 4:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Rand Read - 16:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Rand Read - 32:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Rand Read - 64:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Rand Write - 1:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Rand Write - 4:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Increment - 16:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Increment - 32:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Increment - 64:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Increment - 16:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Increment - 32:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Increment - 64:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Rand Read - 128:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Rand Write - 16:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Rand Write - 32:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Rand Write - 64:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Rand Read - 1:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Rand Read - 4:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Increment - 128:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Rand Read - 1:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Rand Read - 4:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Rand Write - 128:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Rand Read - 16:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Rand Read - 32:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Rand Read - 64:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Rand Write - 1:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Rand Write - 4:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Rand Read - 16:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Rand Read - 32:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Rand Read - 64:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Rand Write - 1:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Rand Write - 4:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Seq Read - 1:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Seq Read - 4:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Rand Read - 128:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Rand Write - 16:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Rand Write - 32:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Rand Read - 128:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Rand Write - 16:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Seq Read - 16:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Seq Read - 32:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Seq Read - 64:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Seq Write - 1:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Seq Write - 4:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Async Rand Read - 1:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Async Rand Read - 4:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Seq Read - 128:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Seq Write - 16:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Seq Write - 32:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Seq Write - 64:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Seq Read - 4:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Seq Read - 1:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Seq Read - 4:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Async Rand Read - 16:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Async Rand Read - 32:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Async Rand Read - 64:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Async Rand Write - 1:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Async Rand Write - 4:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Seq Write - 128:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Seq Read - 16:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Seq Read - 32:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Seq Read - 64:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Seq Write - 1:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Seq Write - 4:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Seq Read - 16:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Seq Read - 32:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Seq Read - 64:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Seq Write - 1:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Seq Write - 4:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Async Rand Read - 128:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Async Rand Write - 16:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Async Rand Write - 32:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Async Rand Write - 64:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Async Rand Read - 1:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Async Rand Read - 4:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Seq Read - 128:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Seq Write - 16:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Seq Write - 32:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Seq Write - 64:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Async Rand Read - 4:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Seq Read - 128:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Seq Write - 16:
    Rows Per Second
    Microseconds - Average Latency
  10000 - Async Rand Write - 128:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Async Rand Read - 16:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Async Rand Read - 64:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Async Rand Write - 1:
    Rows Per Second
    Microseconds - Average Latency
  1000000 - Async Rand Write - 4:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Async Rand Read - 16:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Async Rand Read - 32:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Async Rand Read - 64:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Async Rand Write - 1:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Async Rand Write - 4:
    Rows Per Second
    Microseconds - Average Latency
  2000000 - Async Rand Read - 128:
    Rows Per Second
    Microseconds - Average Latency
BRL-CAD