Hetzner

Compare different Hetzner configuration models

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2203141-ADEP-HETZNER21
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results

Limit displaying results to tests within:

C/C++ Compiler Tests 8 Tests
CPU Massive 12 Tests
Database Test Suite 13 Tests
Go Language Tests 3 Tests
Java Tests 2 Tests
Common Kernel Benchmarks 6 Tests
Multi-Core 4 Tests
Node.js + NPM Tests 2 Tests
Programmer / Developer System Benchmarks 3 Tests
Server 26 Tests
Server CPU Tests 5 Tests
Single-Threaded 5 Tests

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Geometric Means Per-Suite/Category
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
AX41-nvme #1
March 09 2022
 
AX41-nvme #2
March 09 2022
 
Invert Hiding All Results Option
  4 Days, 1 Hour, 1 Minute
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


HetznerOpenBenchmarking.orgPhoronix Test SuiteAMD Ryzen 5 3600 6-Core @ 3.60GHz (6 Cores / 12 Threads)ASUS Pro WS 565-ACE (2424 BIOS)AMD Starship/Matisse2 x 32 GB DDR4-3200MT/s Samsung M378A4G43AB2-CWE2 x 512GB SAMSUNG MZVL2512HCJQ-00B00ASPEEDIntel I210Debian 115.10.0-11-amd64 (x86_64)GCC 10.2.1 20210110ext41024x768ProcessorMotherboardChipsetMemoryDiskGraphicsNetworkOSKernelCompilerFile-SystemScreen ResolutionHetzner BenchmarksSystem Logs- Transparent Huge Pages: always- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-mutex --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-10-Km9U7s/gcc-10-10.2.1/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-10-Km9U7s/gcc-10-10.2.1/debian/tmp-gcn/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - NONE / errors=remount-ro,relatime,rw,stripe=256 / raid0 nvme0n1p3[0] nvme1n1p3[1] Block Size: 4096 - AX41-nvme #1: Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0x8701021 - AX41-nvme #2: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0x8701021 - OpenJDK Runtime Environment (build 11.0.14+9-post-Debian-1deb11u1)- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Vulnerable + spectre_v1: Vulnerable: __user pointer sanitization and usercopy barriers only; no swapgs barriers + spectre_v2: Vulnerable IBPB: disabled STIBP: disabled + srbds: Not affected + tsx_async_abort: Not affected

Hetznerleveldb: Hot Readleveldb: Fill Syncleveldb: Fill Syncleveldb: Overwriteleveldb: Overwriteleveldb: Rand Fillleveldb: Rand Fillleveldb: Rand Readleveldb: Seek Randleveldb: Rand Deleteleveldb: Seq Fillleveldb: Seq Fillsqlite: 1blogbench: Readblogbench: Writesimdjson: Kostyasimdjson: LargeRandsimdjson: PartialTweetssimdjson: DistinctUserIDnode-express-loadtest: ebizzy: perl-benchmark: Pod2htmlperl-benchmark: Interpreteropenssl: SHA256openssl: RSA4096openssl: RSA4096node-web-tooling: couchdb: 100 - 1000 - 24keydb: mysqlslap: 1mysqlslap: 8mysqlslap: 16mysqlslap: 32mysqlslap: 64mysqlslap: 128mysqlslap: 256mysqlslap: 512mysqlslap: 1024mysqlslap: 2048mysqlslap: 4096pgbench: 1 - 1 - Read Onlypgbench: 1 - 1 - Read Only - Average Latencypgbench: 1 - 1 - Read Writepgbench: 1 - 1 - Read Write - Average Latencypgbench: 1 - 50 - Read Onlypgbench: 1 - 50 - Read Only - Average Latencypgbench: 1 - 100 - Read Onlypgbench: 1 - 100 - Read Only - Average Latencypgbench: 1 - 250 - Read Onlypgbench: 1 - 250 - Read Only - Average Latencypgbench: 1 - 50 - Read Writepgbench: 1 - 50 - Read Write - Average Latencypgbench: 1 - 500 - Read Onlypgbench: 1 - 500 - Read Only - Average Latencypgbench: 100 - 1 - Read Onlypgbench: 100 - 1 - Read Only - Average Latencypgbench: 1 - 100 - Read Writepgbench: 1 - 100 - Read Write - Average Latencypgbench: 1 - 250 - Read Writepgbench: 1 - 250 - Read Write - Average Latencypgbench: 1 - 500 - Read Writepgbench: 1 - 500 - Read Write - Average Latencypgbench: 100 - 1 - Read Writepgbench: 100 - 1 - Read Write - Average Latencypgbench: 100 - 50 - Read Onlypgbench: 100 - 50 - Read Only - Average Latencypgbench: 1000 - 1 - Read Onlypgbench: 1000 - 1 - Read Only - Average Latencypgbench: 100 - 100 - Read Onlypgbench: 100 - 100 - Read Only - Average Latencypgbench: 100 - 250 - Read Onlypgbench: 100 - 250 - Read Only - Average Latencypgbench: 100 - 50 - Read Writepgbench: 100 - 50 - Read Write - Average Latencypgbench: 100 - 500 - Read Onlypgbench: 100 - 500 - Read Only - Average Latencypgbench: 1000 - 1 - Read Writepgbench: 1000 - 1 - Read Write - Average Latencypgbench: 1000 - 50 - Read Onlypgbench: 1000 - 50 - Read Only - Average Latencypgbench: 10000 - 1 - Read Onlypgbench: 10000 - 1 - Read Only - Average Latencypgbench: 25000 - 1 - Read Onlypgbench: 25000 - 1 - Read Only - Average Latencypgbench: 100 - 100 - Read Writepgbench: 100 - 100 - Read Write - Average Latencypgbench: 100 - 250 - Read Writepgbench: 100 - 250 - Read Write - Average Latencypgbench: 100 - 500 - Read Writepgbench: 100 - 500 - Read Write - Average Latencypgbench: 1000 - 100 - Read Onlypgbench: 1000 - 100 - Read Only - Average Latencypgbench: 1000 - 250 - Read Onlypgbench: 1000 - 250 - Read Only - Average Latencypgbench: 1000 - 50 - Read Writepgbench: 1000 - 50 - Read Write - Average Latencypgbench: 1000 - 500 - Read Onlypgbench: 1000 - 500 - Read Only - Average Latencypgbench: 10000 - 1 - Read Writepgbench: 10000 - 1 - Read Write - Average Latencypgbench: 10000 - 50 - Read Onlypgbench: 10000 - 50 - Read Only - Average Latencypgbench: 25000 - 1 - Read Writepgbench: 25000 - 1 - Read Write - Average Latencypgbench: 25000 - 50 - Read Onlypgbench: 25000 - 50 - Read Only - Average Latencypgbench: 1000 - 100 - Read Writepgbench: 1000 - 100 - Read Write - Average Latencypgbench: 1000 - 250 - Read Writepgbench: 1000 - 250 - Read Write - Average Latencypgbench: 1000 - 500 - Read Writepgbench: 1000 - 500 - Read Write - Average Latencypgbench: 10000 - 100 - Read Onlypgbench: 10000 - 100 - Read Only - Average Latencypgbench: 10000 - 250 - Read Onlypgbench: 10000 - 250 - Read Only - Average Latencypgbench: 10000 - 50 - Read Writepgbench: 10000 - 50 - Read Write - Average Latencypgbench: 10000 - 500 - Read Onlypgbench: 10000 - 500 - Read Only - Average Latencypgbench: 25000 - 100 - Read Onlypgbench: 25000 - 100 - Read Only - Average Latencypgbench: 25000 - 250 - Read Onlypgbench: 25000 - 250 - Read Only - Average Latencypgbench: 25000 - 50 - Read Writepgbench: 25000 - 50 - Read Write - Average Latencypgbench: 25000 - 500 - Read Onlypgbench: 25000 - 500 - Read Only - Average Latencypgbench: 10000 - 100 - Read Writepgbench: 10000 - 100 - Read Write - Average Latencypgbench: 10000 - 250 - Read Writepgbench: 10000 - 250 - Read Write - Average Latencypgbench: 10000 - 500 - Read Writepgbench: 10000 - 500 - Read Write - Average Latencypgbench: 25000 - 100 - Read Writepgbench: 25000 - 100 - Read Write - Average Latencypgbench: 25000 - 250 - Read Writepgbench: 25000 - 250 - Read Write - Average Latencypgbench: 25000 - 500 - Read Writepgbench: 25000 - 500 - Read Write - Average Latencysqlite-speedtest: Timed Time - Size 1,000redis: LPOPredis: SADDredis: LPUSHredis: GETredis: SETcassandra: Readscassandra: Writescassandra: Mixed 1:1cassandra: Mixed 1:3rocksdb: Rand Fillrocksdb: Rand Readrocksdb: Update Randrocksdb: Seq Fillrocksdb: Rand Fill Syncrocksdb: Read While Writingrocksdb: Read Rand Write Randhbase: Increment - 1hbase: Increment - 1hbase: Increment - 4hbase: Increment - 4hbase: Increment - 16hbase: Increment - 16hbase: Increment - 32hbase: Increment - 32hbase: Increment - 64hbase: Increment - 64hbase: Increment - 128hbase: Increment - 128hbase: Increment - 256hbase: Increment - 256hbase: Rand Read - 1hbase: Rand Read - 1hbase: Rand Read - 4hbase: Rand Read - 4hbase: Rand Read - 16hbase: Rand Read - 16hbase: Rand Read - 32hbase: Rand Read - 32hbase: Rand Read - 64hbase: Rand Read - 64hbase: Rand Write - 1hbase: Rand Write - 1hbase: Rand Write - 4hbase: Rand Write - 4hbase: Rand Read - 128hbase: Rand Read - 128hbase: Rand Read - 256hbase: Rand Read - 256hbase: Rand Write - 16hbase: Rand Write - 16hbase: Rand Write - 32hbase: Rand Write - 32hbase: Rand Write - 64hbase: Rand Write - 64hbase: Rand Write - 128hbase: Rand Write - 128hbase: Rand Write - 256hbase: Rand Write - 256hbase: Seq Read - 1hbase: Seq Read - 1hbase: Seq Read - 4hbase: Seq Read - 4hbase: Seq Read - 16hbase: Seq Read - 16hbase: Seq Read - 32hbase: Seq Read - 32hbase: Seq Read - 64hbase: Seq Read - 64hbase: Seq Write - 1hbase: Seq Write - 1hbase: Seq Write - 4hbase: Seq Write - 4hbase: Async Rand Read - 1hbase: Async Rand Read - 1hbase: Async Rand Read - 4hbase: Async Rand Read - 4hbase: Seq Read - 128hbase: Seq Read - 128hbase: Seq Read - 256hbase: Seq Read - 256hbase: Seq Write - 16hbase: Seq Write - 16hbase: Seq Write - 32hbase: Seq Write - 32hbase: Seq Write - 64hbase: Seq Write - 64hbase: Async Rand Read - 16hbase: Async Rand Read - 16hbase: Async Rand Read - 32hbase: Async Rand Read - 32hbase: Async Rand Read - 64hbase: Async Rand Read - 64hbase: Async Rand Write - 1hbase: Async Rand Write - 1hbase: Async Rand Write - 4hbase: Async Rand Write - 4hbase: Seq Write - 128hbase: Seq Write - 128hbase: Seq Write - 256hbase: Seq Write - 256hbase: Async Rand Read - 128hbase: Async Rand Read - 128hbase: Async Rand Read - 256hbase: Async Rand Read - 256hbase: Async Rand Write - 16hbase: Async Rand Write - 16hbase: Async Rand Write - 32hbase: Async Rand Write - 32hbase: Async Rand Write - 64hbase: Async Rand Write - 64hbase: Async Rand Write - 128hbase: Async Rand Write - 128hbase: Async Rand Write - 256hbase: Async Rand Write - 256apache-siege: 250phpbench: PHP Benchmark Suitephp: Zend benchphp: Zend micro_benchinfluxdb: 4 - 10000 - 2,5000,1 - 10000influxdb: 64 - 10000 - 2,5000,1 - 10000influxdb: 1024 - 10000 - 2,5000,1 - 10000AX41-nvme #1AX41-nvme #25.9020.19913.91945.029.47545.029.4625.9558.99027.91846.428.58862.8422160081155532.740.903.744.2071135535670.119309840.0006867884313116071751.1113279.311.3155.170637796.221131586574534460376323339111106100404360.0256621.5142569750.1942636070.3802359731.06070071.5101826072.738370900.027693144.437636393.360541924.3376011.6692417310.207324860.0312501800.4002445801.02265567.6261649283.0325301.8872014320.24869490.14437700.265768213.116767832.564793563.0192103150.4751870161.337488810.3061783652.8033672.7551387490.3603772.7321129950.442777712.877965725.892915554.6401331280.7521156202.162463710.7871011484.9431145160.873984292.540442411.305878585.691616416.275674337.196666575.242547318.280648138.634644677.88058.9212793879.402154766.251667810.172503451.251869222.526401887205653706522769564436248436384664849987109413740671217254144596832128124536472965269360551789123451022250648963522117511567045056114110139141306228155498414103584925658833138810920151843168224210599219762247176456412166599825200502131818464544870281663922407477243095177674125482729827360187805250095791171871090113815224638764963392559963555661847300721887120372613591043946710422209180346619368328125776421503040323937914374124284892566148477133142398302638504664668775.166389780.3331.6241185073.11342134.21351716.3OpenBenchmarking.org

LevelDB

LevelDB is a key-value storage library developed by Google that supports making use of Snappy for data compression and has other modern features. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMicroseconds Per Op, Fewer Is BetterLevelDB 1.22Benchmark: Hot ReadAX41-nvme #11.3282.6563.9845.3126.64SE +/- 0.058, N = 35.9021. (CXX) g++ options: -O3 -lsnappy -lpthread

OpenBenchmarking.orgMB/s, More Is BetterLevelDB 1.22Benchmark: Fill SyncAX41-nvme #10.02250.0450.06750.090.1125SE +/- 0.00, N = 30.11. (CXX) g++ options: -O3 -lsnappy -lpthread

OpenBenchmarking.orgMicroseconds Per Op, Fewer Is BetterLevelDB 1.22Benchmark: Fill SyncAX41-nvme #12K4K6K8K10KSE +/- 48.67, N = 39913.921. (CXX) g++ options: -O3 -lsnappy -lpthread

OpenBenchmarking.orgMB/s, More Is BetterLevelDB 1.22Benchmark: OverwriteAX41-nvme #11020304050SE +/- 0.10, N = 345.01. (CXX) g++ options: -O3 -lsnappy -lpthread

OpenBenchmarking.orgMicroseconds Per Op, Fewer Is BetterLevelDB 1.22Benchmark: OverwriteAX41-nvme #1714212835SE +/- 0.06, N = 329.481. (CXX) g++ options: -O3 -lsnappy -lpthread

OpenBenchmarking.orgMB/s, More Is BetterLevelDB 1.22Benchmark: Random FillAX41-nvme #11020304050SE +/- 0.03, N = 345.01. (CXX) g++ options: -O3 -lsnappy -lpthread

OpenBenchmarking.orgMicroseconds Per Op, Fewer Is BetterLevelDB 1.22Benchmark: Random FillAX41-nvme #1714212835SE +/- 0.01, N = 329.461. (CXX) g++ options: -O3 -lsnappy -lpthread

OpenBenchmarking.orgMicroseconds Per Op, Fewer Is BetterLevelDB 1.22Benchmark: Random ReadAX41-nvme #11.33992.67984.01975.35966.6995SE +/- 0.041, N = 155.9551. (CXX) g++ options: -O3 -lsnappy -lpthread

OpenBenchmarking.orgMicroseconds Per Op, Fewer Is BetterLevelDB 1.22Benchmark: Seek RandomAX41-nvme #13691215SE +/- 0.031, N = 38.9901. (CXX) g++ options: -O3 -lsnappy -lpthread

OpenBenchmarking.orgMicroseconds Per Op, Fewer Is BetterLevelDB 1.22Benchmark: Random DeleteAX41-nvme #1714212835SE +/- 0.05, N = 327.921. (CXX) g++ options: -O3 -lsnappy -lpthread

OpenBenchmarking.orgMB/s, More Is BetterLevelDB 1.22Benchmark: Sequential FillAX41-nvme #11122334455SE +/- 0.25, N = 346.41. (CXX) g++ options: -O3 -lsnappy -lpthread

OpenBenchmarking.orgMicroseconds Per Op, Fewer Is BetterLevelDB 1.22Benchmark: Sequential FillAX41-nvme #1714212835SE +/- 0.16, N = 328.591. (CXX) g++ options: -O3 -lsnappy -lpthread

SQLite

This is a simple benchmark of SQLite. At present this test profile just measures the time to perform a pre-defined number of insertions on an indexed database. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSQLite 3.30.1Threads / Copies: 1AX41-nvme #11428425670SE +/- 0.51, N = 1562.841. (CC) gcc options: -O2 -lz -lm -ldl -lpthread

BlogBench

BlogBench is designed to replicate the load of a real-world busy file server by stressing the file-system with multiple threads of random reads, writes, and rewrites. The behavior is mimicked of that of a blog by creating blogs with content and pictures, modifying blog posts, adding comments to these blogs, and then reading the content of the blogs. All of these blogs generated are created locally with fake content and pictures. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFinal Score, More Is BetterBlogBench 1.1Test: ReadAX41-nvme #1500K1000K1500K2000K2500KSE +/- 12333.21, N = 321600811. (CC) gcc options: -O2 -pthread

OpenBenchmarking.orgFinal Score, More Is BetterBlogBench 1.1Test: WriteAX41-nvme #13K6K9K12K15KSE +/- 64.79, N = 3155531. (CC) gcc options: -O2 -pthread

simdjson

This is a benchmark of SIMDJSON, a high performance JSON parser. SIMDJSON aims to be the fastest JSON parser and is used by projects like Microsoft FishStore, Yandex ClickHouse, Shopify, and others. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 1.0Throughput Test: KostyaAX41-nvme #10.61651.2331.84952.4663.0825SE +/- 0.01, N = 32.741. (CXX) g++ options: -O3 -pthread

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 1.0Throughput Test: LargeRandomAX41-nvme #10.20250.4050.60750.811.0125SE +/- 0.00, N = 30.901. (CXX) g++ options: -O3 -pthread

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 1.0Throughput Test: PartialTweetsAX41-nvme #10.84151.6832.52453.3664.2075SE +/- 0.01, N = 33.741. (CXX) g++ options: -O3 -pthread

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 1.0Throughput Test: DistinctUserIDAX41-nvme #10.9451.892.8353.784.725SE +/- 0.01, N = 34.201. (CXX) g++ options: -O3 -pthread

Node.js Express HTTP Load Test

A Node.js Express server with a Node-based loadtest client for facilitating HTTP benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterNode.js Express HTTP Load TestAX41-nvme #115003000450060007500SE +/- 73.48, N = 37113

ebizzy

This is a test of ebizzy, a program to generate workloads resembling web server workloads. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRecords/s, More Is Betterebizzy 0.3AX41-nvme #1120K240K360K480K600KSE +/- 9869.98, N = 125535671. (CC) gcc options: -pthread -lpthread -O3 -march=native

Perl Benchmarks

Perl benchmark suite that can be used to compare the relative speed of different versions of perl. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterPerl BenchmarksTest: Pod2htmlAX41-nvme #10.02680.05360.08040.10720.134SE +/- 0.00112073, N = 30.11930984

OpenBenchmarking.orgSeconds, Fewer Is BetterPerl BenchmarksTest: InterpreterAX41-nvme #10.00020.00040.00060.00080.001SE +/- 0.00000288, N = 30.00068678

OpenSSL

OpenSSL is an open-source toolkit that implements SSL (Secure Sockets Layer) and TLS (Transport Layer Security) protocols. This test profile makes use of the built-in "openssl speed" benchmarking capabilities. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgbyte/s, More Is BetterOpenSSL 3.0Algorithm: SHA256AX41-nvme #12000M4000M6000M8000M10000MSE +/- 59968754.96, N = 384313116071. (CC) gcc options: -pthread -m64 -O3 -lssl -lcrypto -ldl

OpenBenchmarking.orgsign/s, More Is BetterOpenSSL 3.0Algorithm: RSA4096AX41-nvme #1400800120016002000SE +/- 0.32, N = 31751.11. (CC) gcc options: -pthread -m64 -O3 -lssl -lcrypto -ldl

OpenBenchmarking.orgverify/s, More Is BetterOpenSSL 3.0Algorithm: RSA4096AX41-nvme #120K40K60K80K100KSE +/- 106.64, N = 3113279.31. (CC) gcc options: -pthread -m64 -O3 -lssl -lcrypto -ldl

Node.js V8 Web Tooling Benchmark

Running the V8 project's Web-Tooling-Benchmark under Node.js. The Web-Tooling-Benchmark stresses JavaScript-related workloads common to web developers like Babel and TypeScript and Babylon. This test profile can test the system's JavaScript performance with Node.js. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgruns/s, More Is BetterNode.js V8 Web Tooling BenchmarkAX41-nvme #13691215SE +/- 0.11, N = 311.31

Apache CouchDB

This is a bulk insertion benchmark of Apache CouchDB. CouchDB is a document-oriented NoSQL database implemented in Erlang. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterApache CouchDB 3.2.1Bulk Size: 100 - Inserts: 1000 - Rounds: 24AX41-nvme #11224364860SE +/- 0.04, N = 355.171. (CXX) g++ options: -flto -lstdc++ -shared -lei

KeyDB

A benchmark of KeyDB as a multi-threaded fork of the Redis server. The KeyDB benchmark is conducted using memtier-benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps/sec, More Is BetterKeyDB 6.2.0AX41-nvme #1140K280K420K560K700KSE +/- 4076.77, N = 3637796.221. (CXX) g++ options: -O2 -levent_openssl -levent -lcrypto -lssl -lpthread -lz -lpcre

MariaDB

This is a MariaDB MySQL database server benchmark making use of mysqlslap. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterMariaDB 10.8.2Clients: 1AX41-nvme #12004006008001000SE +/- 14.56, N = 1211311. (CXX) g++ options: -pie -fPIC -fstack-protector -O3 -pthread -lcrypt -lz -lm -lssl -lcrypto -lpthread -ldl

OpenBenchmarking.orgQueries Per Second, More Is BetterMariaDB 10.8.2Clients: 8AX41-nvme #1130260390520650SE +/- 7.79, N = 35861. (CXX) g++ options: -pie -fPIC -fstack-protector -O3 -pthread -lcrypt -lz -lm -lssl -lcrypto -lpthread -ldl

OpenBenchmarking.orgQueries Per Second, More Is BetterMariaDB 10.8.2Clients: 16AX41-nvme #1120240360480600SE +/- 0.79, N = 35741. (CXX) g++ options: -pie -fPIC -fstack-protector -O3 -pthread -lcrypt -lz -lm -lssl -lcrypto -lpthread -ldl

OpenBenchmarking.orgQueries Per Second, More Is BetterMariaDB 10.8.2Clients: 32AX41-nvme #1120240360480600SE +/- 3.16, N = 35341. (CXX) g++ options: -pie -fPIC -fstack-protector -O3 -pthread -lcrypt -lz -lm -lssl -lcrypto -lpthread -ldl

OpenBenchmarking.orgQueries Per Second, More Is BetterMariaDB 10.8.2Clients: 64AX41-nvme #1100200300400500SE +/- 3.70, N = 34601. (CXX) g++ options: -pie -fPIC -fstack-protector -O3 -pthread -lcrypt -lz -lm -lssl -lcrypto -lpthread -ldl

OpenBenchmarking.orgQueries Per Second, More Is BetterMariaDB 10.8.2Clients: 128AX41-nvme #180160240320400SE +/- 3.73, N = 53761. (CXX) g++ options: -pie -fPIC -fstack-protector -O3 -pthread -lcrypt -lz -lm -lssl -lcrypto -lpthread -ldl

OpenBenchmarking.orgQueries Per Second, More Is BetterMariaDB 10.8.2Clients: 256AX41-nvme #170140210280350SE +/- 2.13, N = 33231. (CXX) g++ options: -pie -fPIC -fstack-protector -O3 -pthread -lcrypt -lz -lm -lssl -lcrypto -lpthread -ldl

OpenBenchmarking.orgQueries Per Second, More Is BetterMariaDB 10.8.2Clients: 512AX41-nvme #170140210280350SE +/- 2.07, N = 33391. (CXX) g++ options: -pie -fPIC -fstack-protector -O3 -pthread -lcrypt -lz -lm -lssl -lcrypto -lpthread -ldl

OpenBenchmarking.orgQueries Per Second, More Is BetterMariaDB 10.8.2Clients: 1024AX41-nvme #120406080100SE +/- 1.00, N = 31111. (CXX) g++ options: -pie -fPIC -fstack-protector -O3 -pthread -lcrypt -lz -lm -lssl -lcrypto -lpthread -ldl

OpenBenchmarking.orgQueries Per Second, More Is BetterMariaDB 10.8.2Clients: 2048AX41-nvme #120406080100SE +/- 0.35, N = 31061. (CXX) g++ options: -pie -fPIC -fstack-protector -O3 -pthread -lcrypt -lz -lm -lssl -lcrypto -lpthread -ldl

OpenBenchmarking.orgQueries Per Second, More Is BetterMariaDB 10.8.2Clients: 4096AX41-nvme #120406080100SE +/- 0.47, N = 31001. (CXX) g++ options: -pie -fPIC -fstack-protector -O3 -pthread -lcrypt -lz -lm -lssl -lcrypto -lpthread -ldl

PostgreSQL pgbench

This is a benchmark of PostgreSQL using pgbench for facilitating the database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 1 - Mode: Read OnlyAX41-nvme #19K18K27K36K45KSE +/- 396.33, N = 3404361. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 1 - Mode: Read Only - Average LatencyAX41-nvme #10.00560.01120.01680.02240.028SE +/- 0.000, N = 30.0251. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 1 - Mode: Read WriteAX41-nvme #1140280420560700SE +/- 8.04, N = 126621. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 1 - Mode: Read Write - Average LatencyAX41-nvme #10.34070.68141.02211.36281.7035SE +/- 0.019, N = 121.5141. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 50 - Mode: Read OnlyAX41-nvme #160K120K180K240K300KSE +/- 1278.07, N = 32569751. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 50 - Mode: Read Only - Average LatencyAX41-nvme #10.04370.08740.13110.17480.2185SE +/- 0.001, N = 30.1941. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 100 - Mode: Read OnlyAX41-nvme #160K120K180K240K300KSE +/- 146.65, N = 32636071. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 100 - Mode: Read Only - Average LatencyAX41-nvme #10.08550.1710.25650.3420.4275SE +/- 0.000, N = 30.3801. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 250 - Mode: Read OnlyAX41-nvme #150K100K150K200K250KSE +/- 1346.96, N = 32359731. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 250 - Mode: Read Only - Average LatencyAX41-nvme #10.23850.4770.71550.9541.1925SE +/- 0.006, N = 31.0601. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 50 - Mode: Read WriteAX41-nvme #1150300450600750SE +/- 8.82, N = 127001. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 50 - Mode: Read Write - Average LatencyAX41-nvme #11632486480SE +/- 0.91, N = 1271.511. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 500 - Mode: Read OnlyAX41-nvme #140K80K120K160K200KSE +/- 988.66, N = 31826071. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 500 - Mode: Read Only - Average LatencyAX41-nvme #10.61611.23221.84832.46443.0805SE +/- 0.015, N = 32.7381. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 1 - Mode: Read OnlyAX41-nvme #18K16K24K32K40KSE +/- 389.76, N = 3370901. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 1 - Mode: Read Only - Average LatencyAX41-nvme #10.00610.01220.01830.02440.0305SE +/- 0.000, N = 30.0271. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 100 - Mode: Read WriteAX41-nvme #1150300450600750SE +/- 7.83, N = 126931. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 100 - Mode: Read Write - Average LatencyAX41-nvme #1306090120150SE +/- 1.65, N = 12144.441. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 250 - Mode: Read WriteAX41-nvme #1140280420560700SE +/- 5.49, N = 126361. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 250 - Mode: Read Write - Average LatencyAX41-nvme #190180270360450SE +/- 3.41, N = 12393.361. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 500 - Mode: Read WriteAX41-nvme #1120240360480600SE +/- 0.35, N = 35411. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1 - Clients: 500 - Mode: Read Write - Average LatencyAX41-nvme #12004006008001000SE +/- 0.60, N = 3924.341. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 1 - Mode: Read WriteAX41-nvme #1130260390520650SE +/- 11.11, N = 126011. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 1 - Mode: Read Write - Average LatencyAX41-nvme #10.37550.7511.12651.5021.8775SE +/- 0.030, N = 121.6691. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 50 - Mode: Read OnlyAX41-nvme #150K100K150K200K250KSE +/- 1014.92, N = 32417311. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 50 - Mode: Read Only - Average LatencyAX41-nvme #10.04660.09320.13980.18640.233SE +/- 0.001, N = 30.2071. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 1 - Mode: Read OnlyAX41-nvme #17K14K21K28K35KSE +/- 102.44, N = 3324861. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 1 - Mode: Read Only - Average LatencyAX41-nvme #10.0070.0140.0210.0280.035SE +/- 0.000, N = 30.0311. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 100 - Mode: Read OnlyAX41-nvme #150K100K150K200K250KSE +/- 798.26, N = 32501801. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 100 - Mode: Read Only - Average LatencyAX41-nvme #10.090.180.270.360.45SE +/- 0.001, N = 30.4001. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 250 - Mode: Read OnlyAX41-nvme #150K100K150K200K250KSE +/- 427.67, N = 32445801. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 250 - Mode: Read Only - Average LatencyAX41-nvme #10.230.460.690.921.15SE +/- 0.002, N = 31.0221. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 50 - Mode: Read WriteAX41-nvme #114002800420056007000SE +/- 5.16, N = 365561. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 50 - Mode: Read Write - Average LatencyAX41-nvme #1246810SE +/- 0.006, N = 37.6261. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 500 - Mode: Read OnlyAX41-nvme #140K80K120K160K200KSE +/- 1102.79, N = 31649281. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 500 - Mode: Read Only - Average LatencyAX41-nvme #10.68221.36442.04662.72883.411SE +/- 0.020, N = 33.0321. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 1 - Mode: Read WriteAX41-nvme #1110220330440550SE +/- 7.03, N = 35301. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 1 - Mode: Read Write - Average LatencyAX41-nvme #10.42460.84921.27381.69842.123SE +/- 0.025, N = 31.8871. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 50 - Mode: Read OnlyAX41-nvme #140K80K120K160K200KSE +/- 757.46, N = 32014321. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 50 - Mode: Read Only - Average LatencyAX41-nvme #10.05580.11160.16740.22320.279SE +/- 0.001, N = 30.2481. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 1 - Mode: Read OnlyAX41-nvme #115003000450060007500SE +/- 76.35, N = 969491. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 1 - Mode: Read Only - Average LatencyAX41-nvme #10.03240.06480.09720.12960.162SE +/- 0.002, N = 90.1441. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 1 - Mode: Read OnlyAX41-nvme #18001600240032004000SE +/- 44.68, N = 337701. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 1 - Mode: Read Only - Average LatencyAX41-nvme #10.05960.11920.17880.23840.298SE +/- 0.003, N = 30.2651. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 100 - Mode: Read WriteAX41-nvme #116003200480064008000SE +/- 198.33, N = 1276821. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 100 - Mode: Read Write - Average LatencyAX41-nvme #13691215SE +/- 0.35, N = 1213.121. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 250 - Mode: Read WriteAX41-nvme #116003200480064008000SE +/- 65.59, N = 376781. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 250 - Mode: Read Write - Average LatencyAX41-nvme #1816243240SE +/- 0.28, N = 332.561. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 500 - Mode: Read WriteAX41-nvme #12K4K6K8K10KSE +/- 70.46, N = 379351. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 100 - Clients: 500 - Mode: Read Write - Average LatencyAX41-nvme #11428425670SE +/- 0.56, N = 363.021. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 100 - Mode: Read OnlyAX41-nvme #150K100K150K200K250KSE +/- 456.72, N = 32103151. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 100 - Mode: Read Only - Average LatencyAX41-nvme #10.10690.21380.32070.42760.5345SE +/- 0.001, N = 30.4751. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 250 - Mode: Read OnlyAX41-nvme #140K80K120K160K200KSE +/- 238.00, N = 31870161. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 250 - Mode: Read Only - Average LatencyAX41-nvme #10.30080.60160.90241.20321.504SE +/- 0.002, N = 31.3371. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 50 - Mode: Read WriteAX41-nvme #110002000300040005000SE +/- 150.94, N = 948881. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 50 - Mode: Read Write - Average LatencyAX41-nvme #13691215SE +/- 0.31, N = 910.311. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 500 - Mode: Read OnlyAX41-nvme #140K80K120K160K200KSE +/- 400.96, N = 31783651. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 500 - Mode: Read Only - Average LatencyAX41-nvme #10.63071.26141.89212.52283.1535SE +/- 0.006, N = 32.8031. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 1 - Mode: Read WriteAX41-nvme #180160240320400SE +/- 13.57, N = 93671. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 1 - Mode: Read Write - Average LatencyAX41-nvme #10.61991.23981.85972.47963.0995SE +/- 0.104, N = 92.7551. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 50 - Mode: Read OnlyAX41-nvme #130K60K90K120K150KSE +/- 93.50, N = 31387491. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 50 - Mode: Read Only - Average LatencyAX41-nvme #10.0810.1620.2430.3240.405SE +/- 0.000, N = 30.3601. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 1 - Mode: Read WriteAX41-nvme #180160240320400SE +/- 23.16, N = 93771. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 1 - Mode: Read Write - Average LatencyAX41-nvme #10.61471.22941.84412.45883.0735SE +/- 0.157, N = 92.7321. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 50 - Mode: Read OnlyAX41-nvme #120K40K60K80K100KSE +/- 1061.49, N = 31129951. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 50 - Mode: Read Only - Average LatencyAX41-nvme #10.09950.1990.29850.3980.4975SE +/- 0.004, N = 30.4421. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 100 - Mode: Read WriteAX41-nvme #117003400510068008500SE +/- 105.67, N = 977771. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 100 - Mode: Read Write - Average LatencyAX41-nvme #13691215SE +/- 0.18, N = 912.881. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 250 - Mode: Read WriteAX41-nvme #12K4K6K8K10KSE +/- 93.61, N = 396571. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 250 - Mode: Read Write - Average LatencyAX41-nvme #1612182430SE +/- 0.25, N = 325.891. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 500 - Mode: Read WriteAX41-nvme #12K4K6K8K10KSE +/- 100.81, N = 591551. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 1000 - Clients: 500 - Mode: Read Write - Average LatencyAX41-nvme #11224364860SE +/- 0.61, N = 554.641. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 100 - Mode: Read OnlyAX41-nvme #130K60K90K120K150KSE +/- 1539.66, N = 41331281. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 100 - Mode: Read Only - Average LatencyAX41-nvme #10.16920.33840.50760.67680.846SE +/- 0.009, N = 40.7521. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 250 - Mode: Read OnlyAX41-nvme #120K40K60K80K100KSE +/- 317.60, N = 31156201. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 250 - Mode: Read Only - Average LatencyAX41-nvme #10.48650.9731.45951.9462.4325SE +/- 0.006, N = 32.1621. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 50 - Mode: Read WriteAX41-nvme #110002000300040005000SE +/- 59.87, N = 346371. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 50 - Mode: Read Write - Average LatencyAX41-nvme #13691215SE +/- 0.14, N = 310.791. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 500 - Mode: Read OnlyAX41-nvme #120K40K60K80K100KSE +/- 304.67, N = 31011481. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 500 - Mode: Read Only - Average LatencyAX41-nvme #11.11222.22443.33664.44885.561SE +/- 0.015, N = 34.9431. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 100 - Mode: Read OnlyAX41-nvme #120K40K60K80K100KSE +/- 869.42, N = 31145161. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 100 - Mode: Read Only - Average LatencyAX41-nvme #10.19640.39280.58920.78560.982SE +/- 0.007, N = 30.8731. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 250 - Mode: Read OnlyAX41-nvme #120K40K60K80K100KSE +/- 382.30, N = 3984291. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 250 - Mode: Read Only - Average LatencyAX41-nvme #10.57151.1431.71452.2862.8575SE +/- 0.010, N = 32.5401. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 50 - Mode: Read WriteAX41-nvme #19001800270036004500SE +/- 53.40, N = 344241. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 50 - Mode: Read Write - Average LatencyAX41-nvme #13691215SE +/- 0.14, N = 311.311. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 500 - Mode: Read OnlyAX41-nvme #120K40K60K80K100KSE +/- 216.72, N = 3878581. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 500 - Mode: Read Only - Average LatencyAX41-nvme #11.28052.5613.84155.1226.4025SE +/- 0.014, N = 35.6911. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 100 - Mode: Read WriteAX41-nvme #113002600390052006500SE +/- 121.09, N = 961641. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 100 - Mode: Read Write - Average LatencyAX41-nvme #148121620SE +/- 0.34, N = 916.281. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 250 - Mode: Read WriteAX41-nvme #114002800420056007000SE +/- 132.28, N = 967431. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 250 - Mode: Read Write - Average LatencyAX41-nvme #1918273645SE +/- 0.75, N = 937.201. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 500 - Mode: Read WriteAX41-nvme #114002800420056007000SE +/- 127.80, N = 966651. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 10000 - Clients: 500 - Mode: Read Write - Average LatencyAX41-nvme #120406080100SE +/- 1.42, N = 975.241. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 100 - Mode: Read WriteAX41-nvme #112002400360048006000SE +/- 83.73, N = 354731. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 100 - Mode: Read Write - Average LatencyAX41-nvme #148121620SE +/- 0.28, N = 318.281. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 250 - Mode: Read WriteAX41-nvme #114002800420056007000SE +/- 87.60, N = 964811. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 250 - Mode: Read Write - Average LatencyAX41-nvme #1918273645SE +/- 0.53, N = 938.631. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgTPS, More Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 500 - Mode: Read WriteAX41-nvme #114002800420056007000SE +/- 292.11, N = 364461. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

OpenBenchmarking.orgms, Fewer Is BetterPostgreSQL pgbench 14.0Scaling Factor: 25000 - Clients: 500 - Mode: Read Write - Average LatencyAX41-nvme #120406080100SE +/- 3.51, N = 377.881. (CC) gcc options: -fno-strict-aliasing -fwrapv -O2 -lpgcommon -lpgport -lpq -lpthread -lrt -ldl -lm

Scaling: Buffer Test - Test: Normal Load - Mode: Read Only

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. E: pgbench: fatal: too many command-line arguments (first is "NORMAL_LOAD")

Scaling: Buffer Test - Test: Normal Load - Mode: Read Write

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. E: pgbench: fatal: too many command-line arguments (first is "NORMAL_LOAD")

Scaling: Buffer Test - Test: Heavy Contention - Mode: Read Only

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. E: pgbench: fatal: too many command-line arguments (first is "HEAVY_CONTENTION")

Scaling: Buffer Test - Test: Heavy Contention - Mode: Read Write

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result. E: pgbench: fatal: too many command-line arguments (first is "HEAVY_CONTENTION")

SQLite Speedtest

This is a benchmark of SQLite's speedtest1 benchmark program with an increased problem size of 1,000. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSQLite Speedtest 3.30Timed Time - Size 1,000AX41-nvme #11326395265SE +/- 0.18, N = 358.921. (CC) gcc options: -O2 -ldl -lz -lpthread

Memtier_benchmark

Memtier_benchmark is a NoSQL Redis/Memcache traffic generation plus benchmarking tool. This current test profile currently just stresses the Redis protocol and basic options exposed wotj a 1:1 Set/Get ratio, 30 pipeline, 100 clients per thread, and thread count equal to the number of CPU cores/threads present. Patches to extend the test are welcome as always. Learn more via the OpenBenchmarking.org test page.

Protocol: Redis

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Redis

Redis is an open-source in-memory data structure store, used as a database, cache, and message broker. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: LPOPAX41-nvme #1600K1200K1800K2400K3000KSE +/- 32206.69, N = 152793879.401. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: SADDAX41-nvme #1500K1000K1500K2000K2500KSE +/- 15091.38, N = 32154766.251. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: LPUSHAX41-nvme #1400K800K1200K1600K2000KSE +/- 16147.13, N = 151667810.171. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: GETAX41-nvme #1500K1000K1500K2000K2500KSE +/- 11640.40, N = 32503451.251. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

OpenBenchmarking.orgRequests Per Second, More Is BetterRedis 6.0.9Test: SETAX41-nvme #1400K800K1200K1600K2000KSE +/- 21488.83, N = 151869222.521. (CXX) g++ options: -MM -MT -g3 -fvisibility=hidden -O3

Apache Cassandra

This is a benchmark of the Apache Cassandra NoSQL database management system making use of cassandra-stress. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 4.0Test: ReadsAX41-nvme #114K28K42K56K70KSE +/- 1175.31, N = 1264018

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 4.0Test: WritesAX41-nvme #120K40K60K80K100KSE +/- 428.19, N = 387205

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 4.0Test: Mixed 1:1AX41-nvme #114K28K42K56K70KSE +/- 278.93, N = 365370

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 4.0Test: Mixed 1:3AX41-nvme #114K28K42K56K70KSE +/- 577.81, N = 1265227

Facebook RocksDB

This is a benchmark of Facebook's RocksDB as an embeddable persistent key-value store for fast storage based on Google's LevelDB. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.22.1Test: Random FillAX41-nvme #1150K300K450K600K750KSE +/- 589.79, N = 36956441. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.22.1Test: Random ReadAX41-nvme #18M16M24M32M40MSE +/- 324948.38, N = 3362484361. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.22.1Test: Update RandomAX41-nvme #180K160K240K320K400KSE +/- 578.74, N = 33846641. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.22.1Test: Sequential FillAX41-nvme #1200K400K600K800K1000KSE +/- 2710.53, N = 38499871. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.22.1Test: Random Fill SyncAX41-nvme #12004006008001000SE +/- 2.03, N = 310941. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.22.1Test: Read While WritingAX41-nvme #1300K600K900K1200K1500KSE +/- 7662.02, N = 313740671. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 6.22.1Test: Read Random Write RandomAX41-nvme #1300K600K900K1200K1500KSE +/- 5651.26, N = 312172541. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

Memcached mcperf

This is a test of twmperf/mcperf with memcached, a distributed memory object caching system. Learn more via the OpenBenchmarking.org test page.

Method: Add - Connections: 1

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Add - Connections: 4

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Get - Connections: 1

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Get - Connections: 4

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Set - Connections: 1

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Set - Connections: 4

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Add - Connections: 16

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Add - Connections: 32

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Add - Connections: 64

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Get - Connections: 16

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Get - Connections: 32

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Get - Connections: 64

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Set - Connections: 16

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Set - Connections: 32

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Set - Connections: 64

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Add - Connections: 128

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Add - Connections: 256

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Get - Connections: 128

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Get - Connections: 256

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Set - Connections: 128

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Set - Connections: 256

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Append - Connections: 1

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Append - Connections: 4

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Delete - Connections: 1

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Delete - Connections: 4

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Append - Connections: 16

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Append - Connections: 32

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Append - Connections: 64

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Delete - Connections: 16

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Delete - Connections: 32

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Delete - Connections: 64

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Prepend - Connections: 1

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Prepend - Connections: 4

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Replace - Connections: 1

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Replace - Connections: 4

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Append - Connections: 128

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Append - Connections: 256

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Delete - Connections: 128

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Delete - Connections: 256

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Prepend - Connections: 16

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Prepend - Connections: 32

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Prepend - Connections: 64

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Replace - Connections: 16

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Replace - Connections: 32

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Replace - Connections: 64

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Prepend - Connections: 128

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Prepend - Connections: 256

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Replace - Connections: 128

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Method: Replace - Connections: 256

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

nginx

This is a benchmark of the lightweight Nginx HTTP(S) web-server. This Nginx web server benchmark test profile makes use of the Golang "Bombardier" program for facilitating the HTTP requests over a fixed period time with a configurable number of concurrent clients. Learn more via the OpenBenchmarking.org test page.

Concurrent Requests: 1

AX41-nvme #1: The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. E: ./nginx: 2: /go/bin/bombardier: not found

Concurrent Requests: 20

AX41-nvme #1: The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. E: ./nginx: 2: /go/bin/bombardier: not found

Concurrent Requests: 100

AX41-nvme #1: The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. E: ./nginx: 2: /go/bin/bombardier: not found

Concurrent Requests: 200

AX41-nvme #1: The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. E: ./nginx: 2: /go/bin/bombardier: not found

Concurrent Requests: 500

AX41-nvme #1: The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. E: ./nginx: 2: /go/bin/bombardier: not found

Concurrent Requests: 1000

AX41-nvme #1: The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. E: ./nginx: 2: /go/bin/bombardier: not found

Apache HBase

This is a benchmark of the Apache HBase non-relational distributed database system inspired from Google's Bigtable. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Increment - Clients: 1AX41-nvme #13K6K9K12K15KSE +/- 119.70, N = 1514459

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Increment - Clients: 1AX41-nvme #11530456075SE +/- 0.57, N = 1568

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Increment - Clients: 4AX41-nvme #17K14K21K28K35KSE +/- 313.89, N = 632128

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Increment - Clients: 4AX41-nvme #1306090120150SE +/- 1.23, N = 6124

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Increment - Clients: 16AX41-nvme #111K22K33K44K55KSE +/- 168.72, N = 353647

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Increment - Clients: 16AX41-nvme #160120180240300SE +/- 1.45, N = 3296

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Increment - Clients: 32AX41-nvme #111K22K33K44K55KSE +/- 397.14, N = 352693

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Increment - Clients: 32AX41-nvme #1130260390520650SE +/- 5.21, N = 3605

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Increment - Clients: 64AX41-nvme #111K22K33K44K55KSE +/- 740.21, N = 351789

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Increment - Clients: 64AX41-nvme #130060090012001500SE +/- 17.69, N = 31234

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Increment - Clients: 128AX41-nvme #111K22K33K44K55KSE +/- 493.77, N = 651022

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Increment - Clients: 128AX41-nvme #15001000150020002500SE +/- 23.56, N = 62506

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Increment - Clients: 256AX41-nvme #110K20K30K40K50KSE +/- 524.05, N = 348963

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Increment - Clients: 256AX41-nvme #111002200330044005500SE +/- 55.43, N = 35221

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Random Read - Clients: 1AX41-nvme #14K8K12K16K20KSE +/- 202.46, N = 1517511

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Random Read - Clients: 1AX41-nvme #11326395265SE +/- 0.71, N = 1556

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Random Read - Clients: 4AX41-nvme #115K30K45K60K75KSE +/- 512.12, N = 1170450

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Random Read - Clients: 4AX41-nvme #11326395265SE +/- 0.44, N = 1156

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Random Read - Clients: 16AX41-nvme #120K40K60K80K100KSE +/- 1345.63, N = 4114110

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Random Read - Clients: 16AX41-nvme #1306090120150SE +/- 1.65, N = 4139

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Random Read - Clients: 32AX41-nvme #130K60K90K120K150KSE +/- 1595.51, N = 3141306

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Random Read - Clients: 32AX41-nvme #150100150200250SE +/- 2.91, N = 3228

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Random Read - Clients: 64AX41-nvme #130K60K90K120K150KSE +/- 586.99, N = 3155498

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Random Read - Clients: 64AX41-nvme #190180270360450SE +/- 1.20, N = 3414

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Random Write - Clients: 1AX41-nvme #120K40K60K80K100KSE +/- 1448.08, N = 15103584

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Random Write - Clients: 1AX41-nvme #13691215SE +/- 0.13, N = 159

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Random Write - Clients: 4AX41-nvme #150K100K150K200K250KSE +/- 29533.20, N = 12256588

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Random Write - Clients: 4AX41-nvme #1816243240SE +/- 16.11, N = 1233

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Random Read - Clients: 128AX41-nvme #130K60K90K120K150KSE +/- 28.92, N = 3138810

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Random Read - Clients: 128AX41-nvme #12004006008001000920

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Random Read - Clients: 256AX41-nvme #130K60K90K120K150KSE +/- 1290.27, N = 3151843

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Random Read - Clients: 256AX41-nvme #1400800120016002000SE +/- 12.86, N = 31682

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Random Write - Clients: 16AX41-nvme #150K100K150K200K250KSE +/- 34033.29, N = 12242105

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Random Write - Clients: 16AX41-nvme #120406080100SE +/- 23.42, N = 1299

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Random Write - Clients: 32AX41-nvme #150K100K150K200K250KSE +/- 32637.23, N = 15219762

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Random Write - Clients: 32AX41-nvme #150100150200250SE +/- 51.07, N = 15247

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Random Write - Clients: 64AX41-nvme #140K80K120K160K200KSE +/- 19730.19, N = 12176456

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Random Write - Clients: 64AX41-nvme #190180270360450SE +/- 42.61, N = 12412

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Random Write - Clients: 128AX41-nvme #140K80K120K160K200KSE +/- 12618.52, N = 12166599

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Random Write - Clients: 128AX41-nvme #12004006008001000SE +/- 73.02, N = 12825

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Random Write - Clients: 256AX41-nvme #140K80K120K160K200KSE +/- 10361.05, N = 9200502

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Random Write - Clients: 256AX41-nvme #130060090012001500SE +/- 74.94, N = 91318

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Sequential Read - Clients: 1AX41-nvme #14K8K12K16K20KSE +/- 455.42, N = 1518464

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Sequential Read - Clients: 1AX41-nvme #11224364860SE +/- 1.38, N = 1554

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Sequential Read - Clients: 4AX41-nvme #110K20K30K40K50KSE +/- 476.58, N = 1248702

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Sequential Read - Clients: 4AX41-nvme #120406080100SE +/- 0.81, N = 1281

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Sequential Read - Clients: 16AX41-nvme #114K28K42K56K70KSE +/- 755.07, N = 366392

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Sequential Read - Clients: 16AX41-nvme #150100150200250SE +/- 2.96, N = 3240

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Sequential Read - Clients: 32AX41-nvme #116K32K48K64K80KSE +/- 291.98, N = 374772

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Sequential Read - Clients: 32AX41-nvme #190180270360450430

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Sequential Read - Clients: 64AX41-nvme #120K40K60K80K100KSE +/- 1702.68, N = 995177

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Sequential Read - Clients: 64AX41-nvme #1150300450600750SE +/- 12.59, N = 9674

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Sequential Write - Clients: 1AX41-nvme #130K60K90K120K150KSE +/- 2137.97, N = 15125482

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Sequential Write - Clients: 1AX41-nvme #1246810SE +/- 0.21, N = 157

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Sequential Write - Clients: 4AX41-nvme #160K120K180K240K300KSE +/- 32788.03, N = 14298273

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Sequential Write - Clients: 4AX41-nvme #11326395265SE +/- 33.44, N = 1460

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Async Random Read - Clients: 1AX41-nvme #14K8K12K16K20KSE +/- 253.90, N = 1518780

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Async Random Read - Clients: 1AX41-nvme #11224364860SE +/- 0.73, N = 1552

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Async Random Read - Clients: 4AX41-nvme #111K22K33K44K55KSE +/- 703.75, N = 1550095

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Async Random Read - Clients: 4AX41-nvme #120406080100SE +/- 1.12, N = 1579

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Sequential Read - Clients: 128AX41-nvme #130K60K90K120K150KSE +/- 1059.41, N = 3117187

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Sequential Read - Clients: 128AX41-nvme #12004006008001000SE +/- 9.53, N = 31090

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Sequential Read - Clients: 256AX41-nvme #120K40K60K80K100KSE +/- 1427.29, N = 3113815

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Sequential Read - Clients: 256AX41-nvme #15001000150020002500SE +/- 27.83, N = 32246

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Sequential Write - Clients: 16AX41-nvme #180K160K240K320K400KSE +/- 25102.50, N = 15387649

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Sequential Write - Clients: 16AX41-nvme #11428425670SE +/- 13.89, N = 1563

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Sequential Write - Clients: 32AX41-nvme #180K160K240K320K400KSE +/- 10954.66, N = 15392559

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Sequential Write - Clients: 32AX41-nvme #120406080100SE +/- 7.31, N = 1596

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Sequential Write - Clients: 64AX41-nvme #180K160K240K320K400KSE +/- 3782.31, N = 15355566

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Sequential Write - Clients: 64AX41-nvme #14080120160200SE +/- 2.33, N = 15184

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Async Random Read - Clients: 16AX41-nvme #116K32K48K64K80KSE +/- 1075.12, N = 1573007

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Async Random Read - Clients: 16AX41-nvme #150100150200250SE +/- 3.53, N = 15218

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Async Random Read - Clients: 32AX41-nvme #120K40K60K80K100KSE +/- 1003.81, N = 1587120

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Async Random Read - Clients: 32AX41-nvme #180160240320400SE +/- 4.35, N = 15372

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Async Random Read - Clients: 64AX41-nvme #113K26K39K52K65KSE +/- 690.46, N = 1261359

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Async Random Read - Clients: 64AX41-nvme #12004006008001000SE +/- 12.35, N = 121043

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Async Random Write - Clients: 1AX41-nvme #12K4K6K8K10KSE +/- 53.01, N = 39467

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Async Random Write - Clients: 1AX41-nvme #120406080100SE +/- 0.67, N = 3104

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Async Random Write - Clients: 4AX41-nvme #15K10K15K20K25KSE +/- 346.31, N = 1522209

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Async Random Write - Clients: 4AX41-nvme #14080120160200SE +/- 3.20, N = 15180

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Sequential Write - Clients: 128AX41-nvme #170K140K210K280K350KSE +/- 3086.82, N = 3346619

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Sequential Write - Clients: 128AX41-nvme #180160240320400SE +/- 3.28, N = 3368

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Sequential Write - Clients: 256AX41-nvme #170K140K210K280K350KSE +/- 761.68, N = 3328125

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Sequential Write - Clients: 256AX41-nvme #12004006008001000SE +/- 1.86, N = 3776

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Async Random Read - Clients: 128AX41-nvme #19K18K27K36K45KSE +/- 711.52, N = 942150

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Async Random Read - Clients: 128AX41-nvme #17001400210028003500SE +/- 57.21, N = 93040

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Async Random Read - Clients: 256AX41-nvme #17K14K21K28K35KSE +/- 544.65, N = 932393

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Async Random Read - Clients: 256AX41-nvme #12K4K6K8K10KSE +/- 142.82, N = 97914

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Async Random Write - Clients: 16AX41-nvme #18K16K24K32K40KSE +/- 696.99, N = 1437412

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Async Random Write - Clients: 16AX41-nvme #190180270360450SE +/- 9.09, N = 14428

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Async Random Write - Clients: 32AX41-nvme #110K20K30K40K50KSE +/- 597.95, N = 348925

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Async Random Write - Clients: 32AX41-nvme #1140280420560700SE +/- 7.94, N = 3661

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Async Random Write - Clients: 64AX41-nvme #110K20K30K40K50KSE +/- 572.62, N = 348477

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Async Random Write - Clients: 64AX41-nvme #130060090012001500SE +/- 15.53, N = 31331

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Async Random Write - Clients: 128AX41-nvme #19K18K27K36K45KSE +/- 726.84, N = 942398

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Async Random Write - Clients: 128AX41-nvme #16001200180024003000SE +/- 53.55, N = 93026

OpenBenchmarking.orgRows Per Second, More Is BetterApache HBase 2.2.3Test: Async Random Write - Clients: 256AX41-nvme #18K16K24K32K40KSE +/- 485.22, N = 338504

OpenBenchmarking.orgMicroseconds - Average Latency, Fewer Is BetterApache HBase 2.2.3Test: Async Random Write - Clients: 256AX41-nvme #114002800420056007000SE +/- 82.55, N = 36646

Apache HTTP Server

This is a test of the Apache HTTPD web server. This Apache HTTPD web server benchmark test profile makes use of the Golang "Bombardier" program for facilitating the HTTP requests over a fixed period time with a configurable number of concurrent clients. Learn more via the OpenBenchmarking.org test page.

Concurrent Requests: 1

AX41-nvme #1: The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. E: ./apache: 2: /go/bin/bombardier: not found

Concurrent Requests: 20

AX41-nvme #1: The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. E: ./apache: 2: /go/bin/bombardier: not found

Concurrent Requests: 100

AX41-nvme #1: The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. E: ./apache: 2: /go/bin/bombardier: not found

Concurrent Requests: 200

AX41-nvme #1: The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. E: ./apache: 2: /go/bin/bombardier: not found

Concurrent Requests: 500

AX41-nvme #1: The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. E: ./apache: 2: /go/bin/bombardier: not found

Concurrent Requests: 1000

AX41-nvme #1: The test quit with a non-zero exit status. The test quit with a non-zero exit status. The test quit with a non-zero exit status. E: ./apache: 2: /go/bin/bombardier: not found

Apache Siege

This is a test of the Apache web server performance being facilitated by the Siege web server benchmark program. Learn more via the OpenBenchmarking.org test page.

Concurrent Users: 200

AX41-nvme #1: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

OpenBenchmarking.orgTransactions Per Second, More Is BetterApache Siege 2.4.29Concurrent Users: 250AX41-nvme #115K30K45K60K75KSE +/- 440.88, N = 368775.161. (CC) gcc options: -O2 -lpthread -ldl -lssl -lcrypto

PHPBench

PHPBench is a benchmark suite for PHP. It performs a large number of simple tests in order to bench various aspects of the PHP interpreter. PHPBench can be used to compare hardware, operating systems, PHP versions, PHP accelerators and caches, compiler options, etc. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgScore, More Is BetterPHPBench 0.8.1PHP Benchmark SuiteAX41-nvme #1140K280K420K560K700KSE +/- 4134.40, N = 3638978

PHP Micro Benchmarks

Various small PHP micro-benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterPHP Micro BenchmarksTest: Zend benchAX41-nvme #10.07490.14980.22470.29960.3745SE +/- 0.002, N = 30.333

OpenBenchmarking.orgSeconds, Fewer Is BetterPHP Micro BenchmarksTest: Zend micro_benchAX41-nvme #10.36540.73081.09621.46161.827SE +/- 0.003, N = 31.624

InfluxDB

This is a benchmark of the InfluxDB open-source time-series database optimized for fast, high-availability storage for IoT and other use-cases. The InfluxDB test profile makes use of InfluxDB Inch for facilitating the benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgval/sec, More Is BetterInfluxDB 1.8.2Concurrent Streams: 4 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 10000AX41-nvme #1300K600K900K1200K1500KSE +/- 4256.05, N = 31185073.1

OpenBenchmarking.orgval/sec, More Is BetterInfluxDB 1.8.2Concurrent Streams: 64 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 10000AX41-nvme #1300K600K900K1200K1500KSE +/- 1882.43, N = 31342134.2

OpenBenchmarking.orgval/sec, More Is BetterInfluxDB 1.8.2Concurrent Streams: 1024 - Batch Size: 10000 - Tags: 2,5000,1 - Points Per Series: 10000AX41-nvme #1300K600K900K1200K1500KSE +/- 1144.06, N = 31351716.3

262 Results Shown

LevelDB:
  Hot Read
  Fill Sync
  Fill Sync
  Overwrite
  Overwrite
  Rand Fill
  Rand Fill
  Rand Read
  Seek Rand
  Rand Delete
  Seq Fill
  Seq Fill
SQLite
BlogBench:
  Read
  Write
simdjson:
  Kostya
  LargeRand
  PartialTweets
  DistinctUserID
Node.js Express HTTP Load Test
ebizzy
Perl Benchmarks:
  Pod2html
  Interpreter
OpenSSL:
  SHA256
  RSA4096
  RSA4096
Node.js V8 Web Tooling Benchmark
Apache CouchDB
KeyDB
MariaDB:
  1
  8
  16
  32
  64
  128
  256
  512
  1024
  2048
  4096
PostgreSQL pgbench:
  1 - 1 - Read Only
  1 - 1 - Read Only - Average Latency
  1 - 1 - Read Write
  1 - 1 - Read Write - Average Latency
  1 - 50 - Read Only
  1 - 50 - Read Only - Average Latency
  1 - 100 - Read Only
  1 - 100 - Read Only - Average Latency
  1 - 250 - Read Only
  1 - 250 - Read Only - Average Latency
  1 - 50 - Read Write
  1 - 50 - Read Write - Average Latency
  1 - 500 - Read Only
  1 - 500 - Read Only - Average Latency
  100 - 1 - Read Only
  100 - 1 - Read Only - Average Latency
  1 - 100 - Read Write
  1 - 100 - Read Write - Average Latency
  1 - 250 - Read Write
  1 - 250 - Read Write - Average Latency
  1 - 500 - Read Write
  1 - 500 - Read Write - Average Latency
  100 - 1 - Read Write
  100 - 1 - Read Write - Average Latency
  100 - 50 - Read Only
  100 - 50 - Read Only - Average Latency
  1000 - 1 - Read Only
  1000 - 1 - Read Only - Average Latency
  100 - 100 - Read Only
  100 - 100 - Read Only - Average Latency
  100 - 250 - Read Only
  100 - 250 - Read Only - Average Latency
  100 - 50 - Read Write
  100 - 50 - Read Write - Average Latency
  100 - 500 - Read Only
  100 - 500 - Read Only - Average Latency
  1000 - 1 - Read Write
  1000 - 1 - Read Write - Average Latency
  1000 - 50 - Read Only
  1000 - 50 - Read Only - Average Latency
  10000 - 1 - Read Only
  10000 - 1 - Read Only - Average Latency
  25000 - 1 - Read Only
  25000 - 1 - Read Only - Average Latency
  100 - 100 - Read Write
  100 - 100 - Read Write - Average Latency
  100 - 250 - Read Write
  100 - 250 - Read Write - Average Latency
  100 - 500 - Read Write
  100 - 500 - Read Write - Average Latency
  1000 - 100 - Read Only
  1000 - 100 - Read Only - Average Latency
  1000 - 250 - Read Only
  1000 - 250 - Read Only - Average Latency
  1000 - 50 - Read Write
  1000 - 50 - Read Write - Average Latency
  1000 - 500 - Read Only
  1000 - 500 - Read Only - Average Latency
  10000 - 1 - Read Write
  10000 - 1 - Read Write - Average Latency
  10000 - 50 - Read Only
  10000 - 50 - Read Only - Average Latency
  25000 - 1 - Read Write
  25000 - 1 - Read Write - Average Latency
  25000 - 50 - Read Only
  25000 - 50 - Read Only - Average Latency
  1000 - 100 - Read Write
  1000 - 100 - Read Write - Average Latency
  1000 - 250 - Read Write
  1000 - 250 - Read Write - Average Latency
  1000 - 500 - Read Write
  1000 - 500 - Read Write - Average Latency
  10000 - 100 - Read Only
  10000 - 100 - Read Only - Average Latency
  10000 - 250 - Read Only
  10000 - 250 - Read Only - Average Latency
  10000 - 50 - Read Write
  10000 - 50 - Read Write - Average Latency
  10000 - 500 - Read Only
  10000 - 500 - Read Only - Average Latency
  25000 - 100 - Read Only
  25000 - 100 - Read Only - Average Latency
  25000 - 250 - Read Only
  25000 - 250 - Read Only - Average Latency
  25000 - 50 - Read Write
  25000 - 50 - Read Write - Average Latency
  25000 - 500 - Read Only
  25000 - 500 - Read Only - Average Latency
  10000 - 100 - Read Write
  10000 - 100 - Read Write - Average Latency
  10000 - 250 - Read Write
  10000 - 250 - Read Write - Average Latency
  10000 - 500 - Read Write
  10000 - 500 - Read Write - Average Latency
  25000 - 100 - Read Write
  25000 - 100 - Read Write - Average Latency
  25000 - 250 - Read Write
  25000 - 250 - Read Write - Average Latency
  25000 - 500 - Read Write
  25000 - 500 - Read Write - Average Latency
SQLite Speedtest
Redis:
  LPOP
  SADD
  LPUSH
  GET
  SET
Apache Cassandra:
  Reads
  Writes
  Mixed 1:1
  Mixed 1:3
Facebook RocksDB:
  Rand Fill
  Rand Read
  Update Rand
  Seq Fill
  Rand Fill Sync
  Read While Writing
  Read Rand Write Rand
Apache HBase:
  Increment - 1:
    Rows Per Second
    Microseconds - Average Latency
  Increment - 4:
    Rows Per Second
    Microseconds - Average Latency
  Increment - 16:
    Rows Per Second
    Microseconds - Average Latency
  Increment - 32:
    Rows Per Second
    Microseconds - Average Latency
  Increment - 64:
    Rows Per Second
    Microseconds - Average Latency
  Increment - 128:
    Rows Per Second
    Microseconds - Average Latency
  Increment - 256:
    Rows Per Second
    Microseconds - Average Latency
  Rand Read - 1:
    Rows Per Second
    Microseconds - Average Latency
  Rand Read - 4:
    Rows Per Second
    Microseconds - Average Latency
  Rand Read - 16:
    Rows Per Second
    Microseconds - Average Latency
  Rand Read - 32:
    Rows Per Second
    Microseconds - Average Latency
  Rand Read - 64:
    Rows Per Second
    Microseconds - Average Latency
  Rand Write - 1:
    Rows Per Second
    Microseconds - Average Latency
  Rand Write - 4:
    Rows Per Second
    Microseconds - Average Latency
  Rand Read - 128:
    Rows Per Second
    Microseconds - Average Latency
  Rand Read - 256:
    Rows Per Second
    Microseconds - Average Latency
  Rand Write - 16:
    Rows Per Second
    Microseconds - Average Latency
  Rand Write - 32:
    Rows Per Second
    Microseconds - Average Latency
  Rand Write - 64:
    Rows Per Second
    Microseconds - Average Latency
  Rand Write - 128:
    Rows Per Second
    Microseconds - Average Latency
  Rand Write - 256:
    Rows Per Second
    Microseconds - Average Latency
  Seq Read - 1:
    Rows Per Second
    Microseconds - Average Latency
  Seq Read - 4:
    Rows Per Second
    Microseconds - Average Latency
  Seq Read - 16:
    Rows Per Second
    Microseconds - Average Latency
  Seq Read - 32:
    Rows Per Second
    Microseconds - Average Latency
  Seq Read - 64:
    Rows Per Second
    Microseconds - Average Latency
  Seq Write - 1:
    Rows Per Second
    Microseconds - Average Latency
  Seq Write - 4:
    Rows Per Second
    Microseconds - Average Latency
  Async Rand Read - 1:
    Rows Per Second
    Microseconds - Average Latency
  Async Rand Read - 4:
    Rows Per Second
    Microseconds - Average Latency
  Seq Read - 128:
    Rows Per Second
    Microseconds - Average Latency
  Seq Read - 256:
    Rows Per Second
    Microseconds - Average Latency
  Seq Write - 16:
    Rows Per Second
    Microseconds - Average Latency
  Seq Write - 32:
    Rows Per Second
    Microseconds - Average Latency
  Seq Write - 64:
    Rows Per Second
    Microseconds - Average Latency
  Async Rand Read - 16:
    Rows Per Second
    Microseconds - Average Latency
  Async Rand Read - 32:
    Rows Per Second
    Microseconds - Average Latency
  Async Rand Read - 64:
    Rows Per Second
    Microseconds - Average Latency
  Async Rand Write - 1:
    Rows Per Second
    Microseconds - Average Latency
  Async Rand Write - 4:
    Rows Per Second
    Microseconds - Average Latency
  Seq Write - 128:
    Rows Per Second
    Microseconds - Average Latency
  Seq Write - 256:
    Rows Per Second
    Microseconds - Average Latency
  Async Rand Read - 128:
    Rows Per Second
    Microseconds - Average Latency
  Async Rand Read - 256:
    Rows Per Second
    Microseconds - Average Latency
  Async Rand Write - 16:
    Rows Per Second
    Microseconds - Average Latency
  Async Rand Write - 32:
    Rows Per Second
    Microseconds - Average Latency
  Async Rand Write - 64:
    Rows Per Second
    Microseconds - Average Latency
  Async Rand Write - 128:
    Rows Per Second
    Microseconds - Average Latency
  Async Rand Write - 256:
    Rows Per Second
    Microseconds - Average Latency
Apache Siege
PHPBench
PHP Micro Benchmarks:
  Zend bench
  Zend micro_bench
InfluxDB:
  4 - 10000 - 2,5000,1 - 10000
  64 - 10000 - 2,5000,1 - 10000
  1024 - 10000 - 2,5000,1 - 10000