r7iz.4xlarge

amazon testing on Amazon Linux 2 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2309085-NE-R7IZ4XLAR65
Jump To Table - Results

Statistics

Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
r7iz.4xlarge
September 08 2023
  2 Days, 2 Hours, 24 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


r7iz.4xlargeOpenBenchmarking.orgPhoronix Test SuiteIntel Xeon Gold 6455B (8 Cores / 16 Threads)Amazon EC2 r7iz.4xlarge (1.0 BIOS)Intel 440FX 82441FX PMC1 x 128 GB 4800MT/s86GB Amazon Elastic Block StoreAmazon ElasticAmazon Linux 25.10.186-179.751.amzn2.x86_64 (x86_64)GCC 10.4.1 20221124xfsamazonProcessorMotherboardChipsetMemoryDiskNetworkOSKernelCompilerFile-SystemSystem LayerR7iz.4xlarge BenchmarksSystem Logs- Transparent Huge Pages: madvise- --build=x86_64-redhat-linux --disable-libunwind-exceptions --enable-__cxa_atexit --enable-bootstrap --enable-checking=release --enable-gnu-indirect-function --enable-gnu-unique-object --enable-initfini-array --enable-languages=c,c++,fortran,lto --enable-multilib --enable-plugin --enable-shared --enable-threads=posix --mandir=/usr/share/man --program-prefix=gcc10- --with-arch_32=x86-64 --with-as=/usr/bin/gcc10-as --with-gcc-major-version-only --with-isl --with-ld=/usr/bin/gcc10-ld --with-linker-hash-style=gnu - CPU Microcode: 0x2b000461- gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Not affected

r7iz.4xlargespark: 40000000 - 2000 - Broadcast Inner Join Test Timespark: 40000000 - 2000 - Inner Join Test Timespark: 40000000 - 2000 - Repartition Test Timespark: 40000000 - 2000 - Group By Test Timespark: 40000000 - 2000 - Calculate Pi Benchmark Using Dataframespark: 40000000 - 2000 - Calculate Pi Benchmarkspark: 40000000 - 2000 - SHA-512 Benchmark Timespark: 40000000 - 1000 - Broadcast Inner Join Test Timespark: 40000000 - 1000 - Inner Join Test Timespark: 40000000 - 1000 - Repartition Test Timespark: 40000000 - 1000 - Group By Test Timespark: 40000000 - 1000 - Calculate Pi Benchmark Using Dataframespark: 40000000 - 1000 - Calculate Pi Benchmarkspark: 40000000 - 1000 - SHA-512 Benchmark Timespark: 20000000 - 2000 - Broadcast Inner Join Test Timespark: 20000000 - 2000 - Inner Join Test Timespark: 20000000 - 2000 - Repartition Test Timespark: 20000000 - 2000 - Group By Test Timespark: 20000000 - 2000 - Calculate Pi Benchmarkspark: 20000000 - 2000 - SHA-512 Benchmark Timespark: 20000000 - 1000 - Broadcast Inner Join Test Timespark: 20000000 - 1000 - Inner Join Test Timespark: 20000000 - 1000 - Repartition Test Timespark: 20000000 - 1000 - Group By Test Timespark: 20000000 - 1000 - Calculate Pi Benchmark Using Dataframespark: 20000000 - 1000 - Calculate Pi Benchmarkspark: 20000000 - 1000 - SHA-512 Benchmark Timespark: 10000000 - 2000 - Broadcast Inner Join Test Timespark: 10000000 - 2000 - Inner Join Test Timespark: 10000000 - 2000 - Repartition Test Timespark: 10000000 - 2000 - Group By Test Timespark: 10000000 - 2000 - Calculate Pi Benchmark Using Dataframespark: 10000000 - 2000 - Calculate Pi Benchmarkspark: 10000000 - 2000 - SHA-512 Benchmark Timespark: 10000000 - 1000 - Broadcast Inner Join Test Timespark: 10000000 - 1000 - Inner Join Test Timespark: 10000000 - 1000 - Repartition Test Timespark: 10000000 - 1000 - Group By Test Timespark: 10000000 - 1000 - Calculate Pi Benchmark Using Dataframespark: 10000000 - 1000 - Calculate Pi Benchmarkspark: 10000000 - 1000 - SHA-512 Benchmark Timespark: 40000000 - 500 - Broadcast Inner Join Test Timespark: 40000000 - 500 - Inner Join Test Timespark: 40000000 - 500 - Repartition Test Timespark: 40000000 - 500 - Group By Test Timespark: 40000000 - 500 - Calculate Pi Benchmarkspark: 40000000 - 500 - SHA-512 Benchmark Timespark: 40000000 - 100 - Broadcast Inner Join Test Timespark: 40000000 - 100 - Inner Join Test Timespark: 40000000 - 100 - Repartition Test Timespark: 40000000 - 100 - Group By Test Timespark: 40000000 - 100 - Calculate Pi Benchmark Using Dataframespark: 40000000 - 100 - Calculate Pi Benchmarkspark: 40000000 - 100 - SHA-512 Benchmark Timespark: 20000000 - 500 - Broadcast Inner Join Test Timespark: 20000000 - 500 - Inner Join Test Timespark: 20000000 - 500 - Repartition Test Timespark: 20000000 - 500 - Group By Test Timespark: 20000000 - 500 - Calculate Pi Benchmarkspark: 20000000 - 500 - SHA-512 Benchmark Timespark: 20000000 - 100 - Broadcast Inner Join Test Timespark: 20000000 - 100 - Inner Join Test Timespark: 20000000 - 100 - Repartition Test Timespark: 20000000 - 100 - Group By Test Timespark: 20000000 - 100 - Calculate Pi Benchmark Using Dataframespark: 20000000 - 100 - Calculate Pi Benchmarkspark: 20000000 - 100 - SHA-512 Benchmark Timespark: 10000000 - 500 - Broadcast Inner Join Test Timespark: 10000000 - 500 - Inner Join Test Timespark: 10000000 - 500 - Repartition Test Timespark: 10000000 - 500 - Group By Test Timespark: 10000000 - 500 - Calculate Pi Benchmarkspark: 10000000 - 500 - SHA-512 Benchmark Timespark: 10000000 - 100 - Broadcast Inner Join Test Timespark: 10000000 - 100 - Inner Join Test Timespark: 10000000 - 100 - Repartition Test Timespark: 10000000 - 100 - Group By Test Timespark: 10000000 - 100 - Calculate Pi Benchmark Using Dataframespark: 10000000 - 100 - Calculate Pi Benchmarkspark: 10000000 - 100 - SHA-512 Benchmark Timespark: 1000000 - 2000 - Broadcast Inner Join Test Timespark: 1000000 - 2000 - Inner Join Test Timespark: 1000000 - 2000 - Repartition Test Timespark: 1000000 - 2000 - Group By Test Timespark: 1000000 - 2000 - Calculate Pi Benchmark Using Dataframespark: 1000000 - 2000 - Calculate Pi Benchmarkspark: 1000000 - 2000 - SHA-512 Benchmark Timespark: 1000000 - 1000 - Broadcast Inner Join Test Timespark: 1000000 - 1000 - Inner Join Test Timespark: 1000000 - 1000 - Repartition Test Timespark: 1000000 - 1000 - Group By Test Timespark: 1000000 - 1000 - Calculate Pi Benchmark Using Dataframespark: 1000000 - 1000 - Calculate Pi Benchmarkspark: 1000000 - 1000 - SHA-512 Benchmark Timespark: 1000000 - 500 - Broadcast Inner Join Test Timespark: 1000000 - 500 - Inner Join Test Timespark: 1000000 - 500 - Repartition Test Timespark: 1000000 - 500 - Group By Test Timespark: 1000000 - 500 - Calculate Pi Benchmarkspark: 1000000 - 500 - SHA-512 Benchmark Timespark: 1000000 - 100 - Broadcast Inner Join Test Timespark: 1000000 - 100 - Inner Join Test Timespark: 1000000 - 100 - Repartition Test Timespark: 1000000 - 100 - Group By Test Timespark: 1000000 - 100 - Calculate Pi Benchmarkspark: 1000000 - 100 - SHA-512 Benchmark Timelibxsmm: 64libxsmm: 32libxsmm: 256libxsmm: 128ncnn: Vulkan GPU - alexnetncnn: Vulkan GPU - FastestDetncnn: Vulkan GPU - vision_transformerncnn: Vulkan GPU - regnety_400mncnn: Vulkan GPU - squeezenet_ssdncnn: Vulkan GPU - yolov4-tinyncnn: Vulkan GPU - resnet50ncnn: Vulkan GPU - vgg16ncnn: Vulkan GPU - googlenetncnn: Vulkan GPU - blazefacencnn: Vulkan GPU - efficientnet-b0ncnn: Vulkan GPU - mnasnetncnn: Vulkan GPU - shufflenet-v2ncnn: Vulkan GPU-v3-v3 - mobilenet-v3ncnn: Vulkan GPU-v2-v2 - mobilenet-v2ncnn: Vulkan GPU - mobilenetncnn: CPU - mnasnetncnn: CPU - FastestDetncnn: CPU - vision_transformerncnn: CPU - regnety_400mncnn: CPU - resnet50ncnn: CPU - alexnetncnn: CPU - vgg16ncnn: CPU - blazefacencnn: CPU - efficientnet-b0ncnn: CPU - shufflenet-v2ncnn: CPU-v3-v3 - mobilenet-v3ncnn: CPU-v2-v2 - mobilenet-v2ncnn: CPU - mobilenetvvenc: Bosphorus 1080p - Fastervvenc: Bosphorus 1080p - Fastvvenc: Bosphorus 4K - Fastervvenc: Bosphorus 4K - Fasttjbench: Decompression Throughputcassandra: Mixed 1:3cassandra: Mixed 1:1cassandra: Writesmemtier-benchmark: Redis - 500 - 1:10memtier-benchmark: Redis - 500 - 10:1memtier-benchmark: Redis - 100 - 1:10memtier-benchmark: Redis - 100 - 10:1memtier-benchmark: Redis - 500 - 5:1memtier-benchmark: Redis - 500 - 1:5memtier-benchmark: Redis - 500 - 1:1memtier-benchmark: Redis - 50 - 1:10memtier-benchmark: Redis - 50 - 10:1memtier-benchmark: Redis - 100 - 5:1memtier-benchmark: Redis - 100 - 1:5memtier-benchmark: Redis - 100 - 1:1memtier-benchmark: Redis - 50 - 5:1memtier-benchmark: Redis - 50 - 1:5memtier-benchmark: Redis - 50 - 1:1nginx: 4000nginx: 1000nginx: 500nginx: 200nginx: 100nginx: 20blogbench: Writeblogbench: Readdacapobench: Tradebeansdacapobench: Tradesoapdacapobench: Eclipsedacapobench: Jythonrenaissance: Genetic Algorithm Using Jenetics + Futuresrenaissance: Akka Unbalanced Cobwebbed Treerenaissance: In-Memory Database Shootoutrenaissance: Finagle HTTP Requestsrenaissance: Apache Spark PageRankrenaissance: Apache Spark Bayesrenaissance: ALS Movie Lensrenaissance: Rand Forestrenaissance: Scala Dottyscimark2: Jacobi Successive Over-Relaxationscimark2: Dense LU Matrix Factorizationscimark2: Sparse Matrix Multiplyscimark2: Fast Fourier Transformscimark2: Monte Carloscimark2: Compositec-ray: Total Time - 4K, 16 Rays Per Pixelstockfish: Total Timesmallpt: Global Illumination Renderer; 128 Samplesgraphics-magick: HWB Color Spacegraphics-magick: Noise-Gaussiangraphics-magick: Resizinggraphics-magick: Enhancedgraphics-magick: Sharpengraphics-magick: Rotategraphics-magick: Swirlcpp-perf-bench: Stepanov Abstractioncpp-perf-bench: Function Objectscpp-perf-bench: Stepanov Vectorcpp-perf-bench: Rand Numberscpp-perf-bench: Math Librarycpp-perf-bench: Ctypecpp-perf-bench: Atolpyperformance: pickle_pure_pythonpyperformance: django_templatepyperformance: python_startuppyperformance: regex_compilepyperformance: crypto_pyaespyperformance: json_loadspyperformance: raytracepyperformance: pathlibpyperformance: nbodypyperformance: floatpyperformance: chaospyperformance: 2to3pyperformance: gox265: Bosphorus 1080px265: Bosphorus 4Kx264: Bosphorus 1080px264: Bosphorus 4Kbotan: ChaCha20Poly1305 - Decryptbotan: ChaCha20Poly1305botan: CAST-256 - Decryptbotan: CAST-256botan: Blowfish - Decryptbotan: Blowfishbotan: Twofish - Decryptbotan: Twofishbotan: AES-256 - Decryptbotan: AES-256botan: KASUMI - Decryptbotan: KASUMIopenssl: ChaCha20-Poly1305openssl: AES-256-GCMopenssl: AES-128-GCMopenssl: ChaCha20openssl: RSA4096openssl: RSA4096openssl: SHA512openssl: SHA256compress-lz4: 9 - Decompression Speedcompress-lz4: 9 - Compression Speedcompress-lz4: 3 - Decompression Speedcompress-lz4: 3 - Compression Speedcompress-lz4: 1 - Decompression Speedcompress-lz4: 1 - Compression Speedcompress-zstd: 19, Long Mode - Decompression Speedcompress-zstd: 19, Long Mode - Compression Speedcompress-zstd: 8, Long Mode - Decompression Speedcompress-zstd: 8, Long Mode - Compression Speedcompress-zstd: 3, Long Mode - Decompression Speedcompress-zstd: 3, Long Mode - Compression Speedcompress-zstd: 19 - Decompression Speedcompress-zstd: 19 - Compression Speedcompress-zstd: 12 - Decompression Speedcompress-zstd: 12 - Compression Speedcompress-zstd: 8 - Decompression Speedcompress-zstd: 8 - Compression Speedcompress-zstd: 3 - Decompression Speedcompress-zstd: 3 - Compression Speedgmpbench: Total Timecachebench: Read / Modify / Writecachebench: Writecachebench: Readintel-mlc: Peak Injection Bandwidth - Stream-Triad Likeintel-mlc: Peak Injection Bandwidth - 1:1 Reads-Writesintel-mlc: Peak Injection Bandwidth - 2:1 Reads-Writesintel-mlc: Peak Injection Bandwidth - 3:1 Reads-Writesintel-mlc: Peak Injection Bandwidth - All Readsintel-mlc: Max Bandwidth - Stream-Triad Likeintel-mlc: Max Bandwidth - 1:1 Reads-Writesintel-mlc: Max Bandwidth - 2:1 Reads-Writesintel-mlc: Max Bandwidth - 3:1 Reads-Writesintel-mlc: Max Bandwidth - All Readsintel-mlc: Idle Latencystream: Addstream: Triadstream: Scalestream: Copyspark: 20000000 - 2000 - Calculate Pi Benchmark Using Dataframespark: 40000000 - 500 - Calculate Pi Benchmark Using Dataframespark: 20000000 - 500 - Calculate Pi Benchmark Using Dataframespark: 10000000 - 500 - Calculate Pi Benchmark Using Dataframespark: 1000000 - 500 - Calculate Pi Benchmark Using Dataframespark: 1000000 - 100 - Calculate Pi Benchmark Using Dataframencnn: Vulkan GPU - resnet18ncnn: CPU - squeezenet_ssdncnn: CPU - yolov4-tinyncnn: CPU - resnet18ncnn: CPU - googlenetdacapobench: H2renaissance: Savina Reactors.IOrenaissance: Apache Spark ALSr7iz.4xlarge27.6829.1923.3021.6711.51219.20354503632.6728.9127.7922.9621.4312.54219.8532.5414.6215.4812.2413.05218.70633683317.2413.9914.5811.8912.6812.63219.1917.118.288.666.706.8211.84219.299.997.878.066.576.5012.24218.6510.2027.3628.1323.1124.56217.8832.2126.7027.4523.1824.3712.23219.9131.8913.5914.2911.9512.69218.8616.7115.2515.3112.4315.4012.18217.9919.347.507.806.396.77218.52036337210.237.908.096.526.7211.64217.59164670510.851.782.271.714.0812.78218.7717590933.441.361.701.523.7811.95220.2039358333.181.151.501.373.57218.2438521213.021.031.411.323.42218.5284735372.94372.6247.0447.5542.53.192.7565.046.075.7913.888.9419.706.290.813.092.202.222.292.428.042.262.8264.596.178.683.1319.980.823.132.252.322.468.1830.11113.6079.2484.338216.3740338158378901909372665199.252377904.812769214.422397346.062301196.122601293.822479211.722805926.482445180.562479007.662736936.782598372.872515318.792845380.042744345.8785555.51100508.87106642.12110863.46110330.3582948.7673331436870312132801363023771732.912120.23442.02363.32443.9908.86272.6567.8621.21379.511709.251190.51539.64190.621001.9057.343274984199.1821072222112523714298150721.72913.72159.029936.696249.80328.69439.09138147.47.3915293.218.941214.596.287.591.526520868.0915.27127.1530.23846.951845.745139.370139.266477.760478.707387.679387.2635505.4395529.15291.05692.7544159103940712750623120316632262328357849221030175992.65961.44137982973116779254678565.757.138559.158.458859.68081.751306.27.781619.5581.21577.6903.21357.514.41569.0219.11637.5552.01538.82305.86370.756544.64734128547.2987954246.067022167090.0173449.5174952.4169108.2147899.5167806.81176001.88176424.12170862.56149243.59113.9128078.9128231.5118927.7171559.612.3614.2715.1214.3314.50967842712.354.316.0414.174.146.3932557237.02531.3OpenBenchmarking.org

Apache Spark

This is a benchmark of Apache Spark with its PySpark interface. Apache Spark is an open-source unified analytics engine for large-scale data processing and dealing with big data. This test profile benchmars the Apache Spark in a single-system configuration using spark-submit. The test makes use of DIYBigData's pyspark-benchmark (https://github.com/DIYBigData/pyspark-benchmark/) for generating of test data and various Apache Spark operations. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 2000 - Broadcast Inner Join Test Timer7iz.4xlarge714212835SE +/- 0.09, N = 327.68

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 2000 - Inner Join Test Timer7iz.4xlarge714212835SE +/- 0.09, N = 329.19

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 2000 - Repartition Test Timer7iz.4xlarge612182430SE +/- 0.06, N = 323.30

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 2000 - Group By Test Timer7iz.4xlarge510152025SE +/- 0.22, N = 321.67

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 2000 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge3691215SE +/- 0.18, N = 311.51

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 2000 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.41, N = 3219.20

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 2000 - SHA-512 Benchmark Timer7iz.4xlarge816243240SE +/- 0.09, N = 332.67

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 1000 - Broadcast Inner Join Test Timer7iz.4xlarge714212835SE +/- 0.05, N = 328.91

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 1000 - Inner Join Test Timer7iz.4xlarge714212835SE +/- 0.10, N = 327.79

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 1000 - Repartition Test Timer7iz.4xlarge612182430SE +/- 0.10, N = 322.96

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 1000 - Group By Test Timer7iz.4xlarge510152025SE +/- 0.08, N = 321.43

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 1000 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge3691215SE +/- 0.34, N = 312.54

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 1000 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.12, N = 3219.85

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 1000 - SHA-512 Benchmark Timer7iz.4xlarge816243240SE +/- 0.12, N = 332.54

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 2000 - Broadcast Inner Join Test Timer7iz.4xlarge48121620SE +/- 0.22, N = 314.62

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 2000 - Inner Join Test Timer7iz.4xlarge48121620SE +/- 0.17, N = 315.48

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 2000 - Repartition Test Timer7iz.4xlarge3691215SE +/- 0.02, N = 312.24

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 2000 - Group By Test Timer7iz.4xlarge3691215SE +/- 0.08, N = 313.05

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 2000 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.10, N = 3218.71

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 2000 - SHA-512 Benchmark Timer7iz.4xlarge48121620SE +/- 0.10, N = 317.24

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 1000 - Broadcast Inner Join Test Timer7iz.4xlarge48121620SE +/- 0.15, N = 313.99

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 1000 - Inner Join Test Timer7iz.4xlarge48121620SE +/- 0.10, N = 314.58

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 1000 - Repartition Test Timer7iz.4xlarge3691215SE +/- 0.02, N = 311.89

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 1000 - Group By Test Timer7iz.4xlarge3691215SE +/- 0.14, N = 312.68

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 1000 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge3691215SE +/- 0.34, N = 312.63

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 1000 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.11, N = 3219.19

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 1000 - SHA-512 Benchmark Timer7iz.4xlarge48121620SE +/- 0.13, N = 317.11

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 2000 - Broadcast Inner Join Test Timer7iz.4xlarge246810SE +/- 0.03, N = 38.28

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 2000 - Inner Join Test Timer7iz.4xlarge246810SE +/- 0.06, N = 38.66

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 2000 - Repartition Test Timer7iz.4xlarge246810SE +/- 0.02, N = 36.70

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 2000 - Group By Test Timer7iz.4xlarge246810SE +/- 0.03, N = 36.82

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 2000 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge3691215SE +/- 0.03, N = 311.84

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 2000 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.29, N = 3219.29

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 2000 - SHA-512 Benchmark Timer7iz.4xlarge3691215SE +/- 0.04, N = 39.99

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 1000 - Broadcast Inner Join Test Timer7iz.4xlarge246810SE +/- 0.04, N = 37.87

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 1000 - Inner Join Test Timer7iz.4xlarge246810SE +/- 0.04, N = 38.06

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 1000 - Repartition Test Timer7iz.4xlarge246810SE +/- 0.08, N = 36.57

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 1000 - Group By Test Timer7iz.4xlarge246810SE +/- 0.07, N = 36.50

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 1000 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge3691215SE +/- 0.29, N = 312.24

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 1000 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.13, N = 3218.65

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 1000 - SHA-512 Benchmark Timer7iz.4xlarge3691215SE +/- 0.10, N = 310.20

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 500 - Broadcast Inner Join Test Timer7iz.4xlarge612182430SE +/- 0.25, N = 327.36

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 500 - Inner Join Test Timer7iz.4xlarge714212835SE +/- 0.20, N = 328.13

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 500 - Repartition Test Timer7iz.4xlarge612182430SE +/- 0.11, N = 323.11

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 500 - Group By Test Timer7iz.4xlarge612182430SE +/- 0.09, N = 324.56

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 500 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.02, N = 3217.88

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 500 - SHA-512 Benchmark Timer7iz.4xlarge714212835SE +/- 0.37, N = 332.21

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 100 - Broadcast Inner Join Test Timer7iz.4xlarge612182430SE +/- 0.06, N = 326.70

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 100 - Inner Join Test Timer7iz.4xlarge612182430SE +/- 0.12, N = 327.45

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 100 - Repartition Test Timer7iz.4xlarge612182430SE +/- 0.01, N = 323.18

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 100 - Group By Test Timer7iz.4xlarge612182430SE +/- 0.35, N = 324.37

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 100 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge3691215SE +/- 0.27, N = 312.23

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 100 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.03, N = 3219.91

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 100 - SHA-512 Benchmark Timer7iz.4xlarge714212835SE +/- 0.34, N = 331.89

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 500 - Broadcast Inner Join Test Timer7iz.4xlarge3691215SE +/- 0.09, N = 313.59

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 500 - Inner Join Test Timer7iz.4xlarge48121620SE +/- 0.20, N = 314.29

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 500 - Repartition Test Timer7iz.4xlarge3691215SE +/- 0.04, N = 311.95

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 500 - Group By Test Timer7iz.4xlarge3691215SE +/- 0.21, N = 312.69

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 500 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.08, N = 3218.86

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 500 - SHA-512 Benchmark Timer7iz.4xlarge48121620SE +/- 0.07, N = 316.71

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 100 - Broadcast Inner Join Test Timer7iz.4xlarge48121620SE +/- 0.03, N = 915.25

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 100 - Inner Join Test Timer7iz.4xlarge48121620SE +/- 0.05, N = 915.31

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 100 - Repartition Test Timer7iz.4xlarge3691215SE +/- 0.04, N = 912.43

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 100 - Group By Test Timer7iz.4xlarge48121620SE +/- 0.11, N = 915.40

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 100 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge3691215SE +/- 0.15, N = 912.18

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 100 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.07, N = 9217.99

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 100 - SHA-512 Benchmark Timer7iz.4xlarge510152025SE +/- 0.16, N = 919.34

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 500 - Broadcast Inner Join Test Timer7iz.4xlarge246810SE +/- 0.10, N = 37.50

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 500 - Inner Join Test Timer7iz.4xlarge246810SE +/- 0.06, N = 37.80

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 500 - Repartition Test Timer7iz.4xlarge246810SE +/- 0.02, N = 36.39

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 500 - Group By Test Timer7iz.4xlarge246810SE +/- 0.02, N = 36.77

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 500 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.05, N = 3218.52

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 500 - SHA-512 Benchmark Timer7iz.4xlarge3691215SE +/- 0.12, N = 310.23

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 100 - Broadcast Inner Join Test Timer7iz.4xlarge246810SE +/- 0.02, N = 37.90

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 100 - Inner Join Test Timer7iz.4xlarge246810SE +/- 0.09, N = 38.09

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 100 - Repartition Test Timer7iz.4xlarge246810SE +/- 0.03, N = 36.52

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 100 - Group By Test Timer7iz.4xlarge246810SE +/- 0.07, N = 36.72

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 100 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge3691215SE +/- 0.14, N = 311.64

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 100 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.13, N = 3217.59

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 100 - SHA-512 Benchmark Timer7iz.4xlarge3691215SE +/- 0.08, N = 310.85

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 2000 - Broadcast Inner Join Test Timer7iz.4xlarge0.40050.8011.20151.6022.0025SE +/- 0.04, N = 31.78

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 2000 - Inner Join Test Timer7iz.4xlarge0.51081.02161.53242.04322.554SE +/- 0.01, N = 32.27

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 2000 - Repartition Test Timer7iz.4xlarge0.38480.76961.15441.53921.924SE +/- 0.03, N = 31.71

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 2000 - Group By Test Timer7iz.4xlarge0.9181.8362.7543.6724.59SE +/- 0.05, N = 34.08

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 2000 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge3691215SE +/- 0.41, N = 312.78

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 2000 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.12, N = 3218.77

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 2000 - SHA-512 Benchmark Timer7iz.4xlarge0.7741.5482.3223.0963.87SE +/- 0.01, N = 33.44

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 1000 - Broadcast Inner Join Test Timer7iz.4xlarge0.3060.6120.9181.2241.53SE +/- 0.02, N = 31.36

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 1000 - Inner Join Test Timer7iz.4xlarge0.38250.7651.14751.531.9125SE +/- 0.02, N = 31.70

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 1000 - Repartition Test Timer7iz.4xlarge0.3420.6841.0261.3681.71SE +/- 0.01, N = 31.52

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 1000 - Group By Test Timer7iz.4xlarge0.85051.7012.55153.4024.2525SE +/- 0.07, N = 33.78

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 1000 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge3691215SE +/- 0.20, N = 311.95

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 1000 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.10, N = 3220.20

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 1000 - SHA-512 Benchmark Timer7iz.4xlarge0.71551.4312.14652.8623.5775SE +/- 0.02, N = 33.18

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 500 - Broadcast Inner Join Test Timer7iz.4xlarge0.25880.51760.77641.03521.294SE +/- 0.02, N = 31.15

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 500 - Inner Join Test Timer7iz.4xlarge0.33750.6751.01251.351.6875SE +/- 0.00, N = 31.50

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 500 - Repartition Test Timer7iz.4xlarge0.30830.61660.92491.23321.5415SE +/- 0.00, N = 31.37

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 500 - Group By Test Timer7iz.4xlarge0.80331.60662.40993.21324.0165SE +/- 0.02, N = 33.57

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 500 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.12, N = 3218.24

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 500 - SHA-512 Benchmark Timer7iz.4xlarge0.67951.3592.03852.7183.3975SE +/- 0.03, N = 33.02

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 100 - Broadcast Inner Join Test Timer7iz.4xlarge0.23180.46360.69540.92721.159SE +/- 0.03, N = 31.03

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 100 - Inner Join Test Timer7iz.4xlarge0.31730.63460.95191.26921.5865SE +/- 0.02, N = 31.41

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 100 - Repartition Test Timer7iz.4xlarge0.2970.5940.8911.1881.485SE +/- 0.01, N = 31.32

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 100 - Group By Test Timer7iz.4xlarge0.76951.5392.30853.0783.8475SE +/- 0.03, N = 33.42

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 100 - Calculate Pi Benchmarkr7iz.4xlarge50100150200250SE +/- 0.05, N = 3218.53

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 100 - SHA-512 Benchmark Timer7iz.4xlarge0.66151.3231.98452.6463.3075SE +/- 0.03, N = 32.94

libxsmm

Libxsmm is an open-source library for specialized dense and sparse matrix operations and deep learning primitives. Libxsmm supports making use of Intel AMX, AVX-512, and other modern CPU instruction set capabilities. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGFLOPS/s, More Is Betterlibxsmm 2-1.17-3645M N K: 64r7iz.4xlarge80160240320400SE +/- 3.03, N = 3372.61. (CXX) g++ options: -dynamic -Bstatic -static-libgcc -l:libgomp.so.1 -lm -lrt -ldl -lstdc++ -pthread -fPIC -std=c++14 -pedantic -O2 -fopenmp -fopenmp-simd -funroll-loops -ftree-vectorize -fdata-sections -ffunction-sections -fvisibility=hidden -march=core-avx2

OpenBenchmarking.orgGFLOPS/s, More Is Betterlibxsmm 2-1.17-3645M N K: 32r7iz.4xlarge50100150200250SE +/- 0.12, N = 3247.01. (CXX) g++ options: -dynamic -Bstatic -static-libgcc -l:libgomp.so.1 -lm -lrt -ldl -lstdc++ -pthread -fPIC -std=c++14 -pedantic -O2 -fopenmp -fopenmp-simd -funroll-loops -ftree-vectorize -fdata-sections -ffunction-sections -fvisibility=hidden -march=core-avx2

OpenBenchmarking.orgGFLOPS/s, More Is Betterlibxsmm 2-1.17-3645M N K: 256r7iz.4xlarge100200300400500SE +/- 2.71, N = 3447.51. (CXX) g++ options: -dynamic -Bstatic -static-libgcc -l:libgomp.so.1 -lm -lrt -ldl -lstdc++ -pthread -fPIC -std=c++14 -pedantic -O2 -fopenmp -fopenmp-simd -funroll-loops -ftree-vectorize -fdata-sections -ffunction-sections -fvisibility=hidden -march=core-avx2

OpenBenchmarking.orgGFLOPS/s, More Is Betterlibxsmm 2-1.17-3645M N K: 128r7iz.4xlarge120240360480600SE +/- 0.21, N = 3542.51. (CXX) g++ options: -dynamic -Bstatic -static-libgcc -l:libgomp.so.1 -lm -lrt -ldl -lstdc++ -pthread -fPIC -std=c++14 -pedantic -O2 -fopenmp -fopenmp-simd -funroll-loops -ftree-vectorize -fdata-sections -ffunction-sections -fvisibility=hidden -march=core-avx2

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: alexnetr7iz.4xlarge0.71781.43562.15342.87123.589SE +/- 0.07, N = 23.19MIN: 2.96 / MAX: 6.911. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: FastestDetr7iz.4xlarge0.61881.23761.85642.47523.094SE +/- 0.02, N = 22.75MIN: 2.64 / MAX: 6.331. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: vision_transformerr7iz.4xlarge1530456075SE +/- 0.21, N = 365.04MIN: 61.57 / MAX: 94.381. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: regnety_400mr7iz.4xlarge246810SE +/- 0.03, N = 36.07MIN: 5.9 / MAX: 9.951. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: squeezenet_ssdr7iz.4xlarge1.30282.60563.90845.21126.514SE +/- 0.12, N = 35.79MIN: 5.53 / MAX: 10.911. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: yolov4-tinyr7iz.4xlarge48121620SE +/- 0.16, N = 313.88MIN: 13 / MAX: 20.721. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: resnet50r7iz.4xlarge246810SE +/- 0.26, N = 38.94MIN: 8.05 / MAX: 15.11. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: vgg16r7iz.4xlarge510152025SE +/- 0.17, N = 319.70MIN: 18.82 / MAX: 26.381. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: googlenetr7iz.4xlarge246810SE +/- 0.08, N = 36.29MIN: 5.88 / MAX: 11.831. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: blazefacer7iz.4xlarge0.18230.36460.54690.72920.9115SE +/- 0.01, N = 30.81MIN: 0.78 / MAX: 0.941. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: efficientnet-b0r7iz.4xlarge0.69531.39062.08592.78123.4765SE +/- 0.01, N = 33.09MIN: 2.96 / MAX: 6.851. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: mnasnetr7iz.4xlarge0.4950.991.4851.982.475SE +/- 0.01, N = 22.20MIN: 2.12 / MAX: 2.761. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: shufflenet-v2r7iz.4xlarge0.49950.9991.49851.9982.4975SE +/- 0.01, N = 32.22MIN: 2.13 / MAX: 5.881. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU-v3-v3 - Model: mobilenet-v3r7iz.4xlarge0.51531.03061.54592.06122.5765SE +/- 0.01, N = 32.29MIN: 2.19 / MAX: 5.951. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU-v2-v2 - Model: mobilenet-v2r7iz.4xlarge0.54451.0891.63352.1782.7225SE +/- 0.03, N = 32.42MIN: 2.31 / MAX: 5.991. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: mobilenetr7iz.4xlarge246810SE +/- 0.03, N = 38.04MIN: 7.81 / MAX: 12.351. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: mnasnetr7iz.4xlarge0.50851.0171.52552.0342.5425SE +/- 0.00, N = 52.26MIN: 2.16 / MAX: 61. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: FastestDetr7iz.4xlarge0.63451.2691.90352.5383.1725SE +/- 0.06, N = 82.82MIN: 2.57 / MAX: 6.671. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: vision_transformerr7iz.4xlarge1428425670SE +/- 0.24, N = 864.59MIN: 61.22 / MAX: 106.251. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: regnety_400mr7iz.4xlarge246810SE +/- 0.03, N = 86.17MIN: 5.97 / MAX: 12.731. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: resnet50r7iz.4xlarge246810SE +/- 0.09, N = 88.68MIN: 8.07 / MAX: 15.551. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: alexnetr7iz.4xlarge0.70431.40862.11292.81723.5215SE +/- 0.02, N = 73.13MIN: 2.95 / MAX: 7.741. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: vgg16r7iz.4xlarge510152025SE +/- 0.23, N = 819.98MIN: 18.91 / MAX: 26.881. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: blazefacer7iz.4xlarge0.18450.3690.55350.7380.9225SE +/- 0.00, N = 80.82MIN: 0.79 / MAX: 4.451. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: efficientnet-b0r7iz.4xlarge0.70431.40862.11292.81723.5215SE +/- 0.01, N = 83.13MIN: 3 / MAX: 7.091. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: shufflenet-v2r7iz.4xlarge0.50631.01261.51892.02522.5315SE +/- 0.01, N = 82.25MIN: 2.15 / MAX: 5.891. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU-v3-v3 - Model: mobilenet-v3r7iz.4xlarge0.5221.0441.5662.0882.61SE +/- 0.01, N = 82.32MIN: 2.22 / MAX: 7.611. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU-v2-v2 - Model: mobilenet-v2r7iz.4xlarge0.55351.1071.66052.2142.7675SE +/- 0.01, N = 82.46MIN: 2.35 / MAX: 6.291. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: mobilenetr7iz.4xlarge246810SE +/- 0.07, N = 88.18MIN: 7.81 / MAX: 13.071. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

VVenC

VVenC is the Fraunhofer Versatile Video Encoder as a fast/efficient H.266/VVC encoder. The vvenc encoder makes use of SIMD Everywhere (SIMDe). The vvenc software is published under the Clear BSD License. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.9Video Input: Bosphorus 1080p - Video Preset: Fasterr7iz.4xlarge714212835SE +/- 0.04, N = 330.111. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.9Video Input: Bosphorus 1080p - Video Preset: Fastr7iz.4xlarge3691215SE +/- 0.03, N = 313.611. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.9Video Input: Bosphorus 4K - Video Preset: Fasterr7iz.4xlarge3691215SE +/- 0.010, N = 39.2481. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects -lpthread

OpenBenchmarking.orgFrames Per Second, More Is BetterVVenC 1.9Video Input: Bosphorus 4K - Video Preset: Fastr7iz.4xlarge0.97611.95222.92833.90444.8805SE +/- 0.005, N = 34.3381. (CXX) g++ options: -O3 -flto=auto -fno-fat-lto-objects -lpthread

libjpeg-turbo tjbench

tjbench is a JPEG decompression/compression benchmark that is part of libjpeg-turbo, a JPEG image codec library optimized for SIMD instructions on modern CPU architectures. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMegapixels/sec, More Is Betterlibjpeg-turbo tjbench 2.1.0Test: Decompression Throughputr7iz.4xlarge50100150200250SE +/- 0.06, N = 3216.371. (CC) gcc options: -O3 -rdynamic

Apache Cassandra

This is a benchmark of the Apache Cassandra NoSQL database management system making use of cassandra-stress. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 4.1.3Test: Mixed 1:3r7iz.4xlarge20K40K60K80K100KSE +/- 554.99, N = 381583

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 4.1.3Test: Mixed 1:1r7iz.4xlarge20K40K60K80K100KSE +/- 308.50, N = 378901

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 4.1.3Test: Writesr7iz.4xlarge20K40K60K80K100KSE +/- 35.17, N = 390937

Redis 7.0.12 + memtier_benchmark

Memtier_benchmark is a NoSQL Redis/Memcache traffic generation plus benchmarking tool developed by Redis Labs. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 500 - Set To Get Ratio: 1:10r7iz.4xlarge600K1200K1800K2400K3000KSE +/- 19928.28, N = 102665199.251. (CXX) g++ options: -O2 -levent_openssl -levent -lssl -lcrypto -lpthread -lz -lpcre

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 500 - Set To Get Ratio: 10:1r7iz.4xlarge500K1000K1500K2000K2500KSE +/- 13861.62, N = 32377904.811. (CXX) g++ options: -O2 -levent_openssl -levent -lssl -lcrypto -lpthread -lz -lpcre

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 100 - Set To Get Ratio: 1:10r7iz.4xlarge600K1200K1800K2400K3000KSE +/- 32773.55, N = 32769214.421. (CXX) g++ options: -O2 -levent_openssl -levent -lssl -lcrypto -lpthread -lz -lpcre

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 100 - Set To Get Ratio: 10:1r7iz.4xlarge500K1000K1500K2000K2500KSE +/- 8558.69, N = 32397346.061. (CXX) g++ options: -O2 -levent_openssl -levent -lssl -lcrypto -lpthread -lz -lpcre

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 500 - Set To Get Ratio: 5:1r7iz.4xlarge500K1000K1500K2000K2500KSE +/- 840.99, N = 32301196.121. (CXX) g++ options: -O2 -levent_openssl -levent -lssl -lcrypto -lpthread -lz -lpcre

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 500 - Set To Get Ratio: 1:5r7iz.4xlarge600K1200K1800K2400K3000KSE +/- 10861.52, N = 32601293.821. (CXX) g++ options: -O2 -levent_openssl -levent -lssl -lcrypto -lpthread -lz -lpcre

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 500 - Set To Get Ratio: 1:1r7iz.4xlarge500K1000K1500K2000K2500KSE +/- 29522.75, N = 32479211.721. (CXX) g++ options: -O2 -levent_openssl -levent -lssl -lcrypto -lpthread -lz -lpcre

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 50 - Set To Get Ratio: 1:10r7iz.4xlarge600K1200K1800K2400K3000KSE +/- 1677.07, N = 32805926.481. (CXX) g++ options: -O2 -levent_openssl -levent -lssl -lcrypto -lpthread -lz -lpcre

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 50 - Set To Get Ratio: 10:1r7iz.4xlarge500K1000K1500K2000K2500KSE +/- 7136.13, N = 32445180.561. (CXX) g++ options: -O2 -levent_openssl -levent -lssl -lcrypto -lpthread -lz -lpcre

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 100 - Set To Get Ratio: 5:1r7iz.4xlarge500K1000K1500K2000K2500KSE +/- 9062.40, N = 32479007.661. (CXX) g++ options: -O2 -levent_openssl -levent -lssl -lcrypto -lpthread -lz -lpcre

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 100 - Set To Get Ratio: 1:5r7iz.4xlarge600K1200K1800K2400K3000KSE +/- 16363.16, N = 32736936.781. (CXX) g++ options: -O2 -levent_openssl -levent -lssl -lcrypto -lpthread -lz -lpcre

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 100 - Set To Get Ratio: 1:1r7iz.4xlarge600K1200K1800K2400K3000KSE +/- 2486.22, N = 32598372.871. (CXX) g++ options: -O2 -levent_openssl -levent -lssl -lcrypto -lpthread -lz -lpcre

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 50 - Set To Get Ratio: 5:1r7iz.4xlarge500K1000K1500K2000K2500KSE +/- 7795.11, N = 32515318.791. (CXX) g++ options: -O2 -levent_openssl -levent -lssl -lcrypto -lpthread -lz -lpcre

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 50 - Set To Get Ratio: 1:5r7iz.4xlarge600K1200K1800K2400K3000KSE +/- 7568.76, N = 32845380.041. (CXX) g++ options: -O2 -levent_openssl -levent -lssl -lcrypto -lpthread -lz -lpcre

OpenBenchmarking.orgOps/sec, More Is BetterRedis 7.0.12 + memtier_benchmark 2.0Protocol: Redis - Clients: 50 - Set To Get Ratio: 1:1r7iz.4xlarge600K1200K1800K2400K3000KSE +/- 18457.67, N = 32744345.871. (CXX) g++ options: -O2 -levent_openssl -levent -lssl -lcrypto -lpthread -lz -lpcre

nginx

This is a benchmark of the lightweight Nginx HTTP(S) web-server. This Nginx web server benchmark test profile makes use of the wrk program for facilitating the HTTP requests over a fixed period time with a configurable number of concurrent clients/connections. HTTPS with a self-signed OpenSSL certificate is used by this test for local benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgRequests Per Second, More Is Betternginx 1.23.2Connections: 4000r7iz.4xlarge20K40K60K80K100KSE +/- 212.42, N = 385555.511. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2

OpenBenchmarking.orgRequests Per Second, More Is Betternginx 1.23.2Connections: 1000r7iz.4xlarge20K40K60K80K100KSE +/- 247.42, N = 3100508.871. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2

OpenBenchmarking.orgRequests Per Second, More Is Betternginx 1.23.2Connections: 500r7iz.4xlarge20K40K60K80K100KSE +/- 109.28, N = 3106642.121. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2

OpenBenchmarking.orgRequests Per Second, More Is Betternginx 1.23.2Connections: 200r7iz.4xlarge20K40K60K80K100KSE +/- 66.67, N = 3110863.461. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2

OpenBenchmarking.orgRequests Per Second, More Is Betternginx 1.23.2Connections: 100r7iz.4xlarge20K40K60K80K100KSE +/- 18.96, N = 3110330.351. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2

OpenBenchmarking.orgRequests Per Second, More Is Betternginx 1.23.2Connections: 20r7iz.4xlarge20K40K60K80K100KSE +/- 107.15, N = 382948.761. (CC) gcc options: -lluajit-5.1 -lm -lssl -lcrypto -lpthread -ldl -std=c99 -O2

BlogBench

BlogBench is designed to replicate the load of a real-world busy file server by stressing the file-system with multiple threads of random reads, writes, and rewrites. The behavior is mimicked of that of a blog by creating blogs with content and pictures, modifying blog posts, adding comments to these blogs, and then reading the content of the blogs. All of these blogs generated are created locally with fake content and pictures. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFinal Score, More Is BetterBlogBench 1.1Test: Writer7iz.4xlarge16003200480064008000SE +/- 16.86, N = 373331. (CC) gcc options: -O2 -pthread

OpenBenchmarking.orgFinal Score, More Is BetterBlogBench 1.1Test: Readr7iz.4xlarge300K600K900K1200K1500KSE +/- 27182.39, N = 914368701. (CC) gcc options: -O2 -pthread

DaCapo Benchmark

This test runs the DaCapo Benchmarks written in Java and intended to test system/CPU performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgmsec, Fewer Is BetterDaCapo Benchmark 9.12-MR1Java Test: Tradebeansr7iz.4xlarge7001400210028003500SE +/- 37.88, N = 43121

OpenBenchmarking.orgmsec, Fewer Is BetterDaCapo Benchmark 9.12-MR1Java Test: Tradesoapr7iz.4xlarge7001400210028003500SE +/- 26.58, N = 203280

OpenBenchmarking.orgmsec, Fewer Is BetterDaCapo Benchmark 9.12-MR1Java Test: Eclipser7iz.4xlarge3K6K9K12K15KSE +/- 50.73, N = 413630

OpenBenchmarking.orgmsec, Fewer Is BetterDaCapo Benchmark 9.12-MR1Java Test: Jythonr7iz.4xlarge5001000150020002500SE +/- 23.16, N = 42377

Renaissance

Renaissance is a suite of benchmarks designed to test the Java JVM from Apache Spark to a Twitter-like service to Scala and other features. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterRenaissance 0.14Test: Genetic Algorithm Using Jenetics + Futuresr7iz.4xlarge400800120016002000SE +/- 11.12, N = 31732.9MIN: 1678.98 / MAX: 1794.4

OpenBenchmarking.orgms, Fewer Is BetterRenaissance 0.14Test: Akka Unbalanced Cobwebbed Treer7iz.4xlarge3K6K9K12K15KSE +/- 138.88, N = 312120.2MIN: 9019.77 / MAX: 12262.45

OpenBenchmarking.orgms, Fewer Is BetterRenaissance 0.14Test: In-Memory Database Shootoutr7iz.4xlarge7001400210028003500SE +/- 26.79, N = 33442.0MIN: 3122.7 / MAX: 3703.32

OpenBenchmarking.orgms, Fewer Is BetterRenaissance 0.14Test: Finagle HTTP Requestsr7iz.4xlarge5001000150020002500SE +/- 6.57, N = 32363.3MIN: 2175.24 / MAX: 2527.83

OpenBenchmarking.orgms, Fewer Is BetterRenaissance 0.14Test: Apache Spark PageRankr7iz.4xlarge5001000150020002500SE +/- 11.32, N = 32443.9MIN: 2193.37 / MAX: 2460.95

OpenBenchmarking.orgms, Fewer Is BetterRenaissance 0.14Test: Apache Spark Bayesr7iz.4xlarge2004006008001000SE +/- 9.21, N = 5908.8MIN: 617.76 / MAX: 940.63

OpenBenchmarking.orgms, Fewer Is BetterRenaissance 0.14Test: ALS Movie Lensr7iz.4xlarge13002600390052006500SE +/- 31.05, N = 36272.6MIN: 6180.94 / MAX: 6867.53

OpenBenchmarking.orgms, Fewer Is BetterRenaissance 0.14Test: Random Forestr7iz.4xlarge120240360480600SE +/- 3.42, N = 3567.8MIN: 510.04 / MAX: 725.19

OpenBenchmarking.orgms, Fewer Is BetterRenaissance 0.14Test: Scala Dottyr7iz.4xlarge130260390520650SE +/- 0.39, N = 3621.2MIN: 518.77 / MAX: 1076.18

SciMark

This test runs the ANSI C version of SciMark 2.0, which is a benchmark for scientific and numerical computing developed by programmers at the National Institute of Standards and Technology. This test is made up of Fast Foruier Transform, Jacobi Successive Over-relaxation, Monte Carlo, Sparse Matrix Multiply, and dense LU matrix factorization benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMflops, More Is BetterSciMark 2.0Computational Test: Jacobi Successive Over-Relaxationr7iz.4xlarge30060090012001500SE +/- 0.08, N = 31379.511. (CC) gcc options: -lm

OpenBenchmarking.orgMflops, More Is BetterSciMark 2.0Computational Test: Dense LU Matrix Factorizationr7iz.4xlarge400800120016002000SE +/- 1.91, N = 31709.251. (CC) gcc options: -lm

OpenBenchmarking.orgMflops, More Is BetterSciMark 2.0Computational Test: Sparse Matrix Multiplyr7iz.4xlarge30060090012001500SE +/- 7.74, N = 31190.511. (CC) gcc options: -lm

OpenBenchmarking.orgMflops, More Is BetterSciMark 2.0Computational Test: Fast Fourier Transformr7iz.4xlarge120240360480600SE +/- 0.52, N = 3539.641. (CC) gcc options: -lm

OpenBenchmarking.orgMflops, More Is BetterSciMark 2.0Computational Test: Monte Carlor7iz.4xlarge4080120160200SE +/- 0.18, N = 3190.621. (CC) gcc options: -lm

OpenBenchmarking.orgMflops, More Is BetterSciMark 2.0Computational Test: Compositer7iz.4xlarge2004006008001000SE +/- 2.01, N = 31001.901. (CC) gcc options: -lm

C-Ray

This is a test of C-Ray, a simple raytracer designed to test the floating-point CPU performance. This test is multi-threaded (16 threads per core), will shoot 8 rays per pixel for anti-aliasing, and will generate a 1600 x 1200 image. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterC-Ray 1.1Total Time - 4K, 16 Rays Per Pixelr7iz.4xlarge1326395265SE +/- 0.04, N = 357.341. (CC) gcc options: -lm -lpthread -O3

Stockfish

This is a test of Stockfish, an advanced open-source C++11 chess benchmark that can scale up to 512 CPU threads. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgNodes Per Second, More Is BetterStockfish 15Total Timer7iz.4xlarge6M12M18M24M30MSE +/- 235411.39, N = 15274984191. (CXX) g++ options: -lgcov -m64 -lpthread -fno-exceptions -std=c++17 -fno-peel-loops -fno-tracer -pedantic -O3 -msse -msse3 -mpopcnt -mavx2 -mavx512f -mavx512bw -mavx512vnni -mavx512dq -mavx512vl -msse4.1 -mssse3 -msse2 -mbmi2 -flto -flto=jobserver

Smallpt

Smallpt is a C++ global illumination renderer written in less than 100 lines of code. Global illumination is done via unbiased Monte Carlo path tracing and there is multi-threading support via the OpenMP library. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSmallpt 1.0Global Illumination Renderer; 128 Samplesr7iz.4xlarge3691215SE +/- 0.004, N = 39.1821. (CXX) g++ options: -fopenmp -O3

GraphicsMagick

This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: HWB Color Spacer7iz.4xlarge200400600800100010721. (CC) gcc options: -fopenmp -O2 -pthread -ljpeg -lXext -lSM -lICE -lX11 -lz -lm -lpthread

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: Noise-Gaussianr7iz.4xlarge501001502002502221. (CC) gcc options: -fopenmp -O2 -pthread -ljpeg -lXext -lSM -lICE -lX11 -lz -lm -lpthread

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: Resizingr7iz.4xlarge2004006008001000SE +/- 2.03, N = 311251. (CC) gcc options: -fopenmp -O2 -pthread -ljpeg -lXext -lSM -lICE -lX11 -lz -lm -lpthread

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: Enhancedr7iz.4xlarge501001502002502371. (CC) gcc options: -fopenmp -O2 -pthread -ljpeg -lXext -lSM -lICE -lX11 -lz -lm -lpthread

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: Sharpenr7iz.4xlarge3060901201501421. (CC) gcc options: -fopenmp -O2 -pthread -ljpeg -lXext -lSM -lICE -lX11 -lz -lm -lpthread

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: Rotater7iz.4xlarge20040060080010009811. (CC) gcc options: -fopenmp -O2 -pthread -ljpeg -lXext -lSM -lICE -lX11 -lz -lm -lpthread

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.38Operation: Swirlr7iz.4xlarge1102203304405505071. (CC) gcc options: -fopenmp -O2 -pthread -ljpeg -lXext -lSM -lICE -lX11 -lz -lm -lpthread

CppPerformanceBenchmarks

CppPerformanceBenchmarks is a set of C++ compiler performance benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterCppPerformanceBenchmarks 9Test: Stepanov Abstractionr7iz.4xlarge510152025SE +/- 0.02, N = 321.731. (CXX) g++ options: -std=c++11 -O3

OpenBenchmarking.orgSeconds, Fewer Is BetterCppPerformanceBenchmarks 9Test: Function Objectsr7iz.4xlarge48121620SE +/- 0.00, N = 313.721. (CXX) g++ options: -std=c++11 -O3

OpenBenchmarking.orgSeconds, Fewer Is BetterCppPerformanceBenchmarks 9Test: Stepanov Vectorr7iz.4xlarge1326395265SE +/- 0.09, N = 359.031. (CXX) g++ options: -std=c++11 -O3

OpenBenchmarking.orgSeconds, Fewer Is BetterCppPerformanceBenchmarks 9Test: Random Numbersr7iz.4xlarge2004006008001000SE +/- 4.12, N = 3936.701. (CXX) g++ options: -std=c++11 -O3

OpenBenchmarking.orgSeconds, Fewer Is BetterCppPerformanceBenchmarks 9Test: Math Libraryr7iz.4xlarge50100150200250SE +/- 0.02, N = 3249.801. (CXX) g++ options: -std=c++11 -O3

OpenBenchmarking.orgSeconds, Fewer Is BetterCppPerformanceBenchmarks 9Test: Ctyper7iz.4xlarge714212835SE +/- 0.09, N = 328.691. (CXX) g++ options: -std=c++11 -O3

OpenBenchmarking.orgSeconds, Fewer Is BetterCppPerformanceBenchmarks 9Test: Atolr7iz.4xlarge918273645SE +/- 0.02, N = 339.091. (CXX) g++ options: -std=c++11 -O3

PyPerformance

PyPerformance is the reference Python performance benchmark suite. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: pickle_pure_pythonr7iz.4xlarge80160240320400381

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: django_templater7iz.4xlarge1122334455SE +/- 0.06, N = 347.4

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: python_startupr7iz.4xlarge246810SE +/- 0.00, N = 37.39

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: regex_compiler7iz.4xlarge306090120150152

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: crypto_pyaesr7iz.4xlarge20406080100SE +/- 0.06, N = 393.2

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: json_loadsr7iz.4xlarge510152025SE +/- 0.00, N = 318.9

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: raytracer7iz.4xlarge90180270360450SE +/- 0.67, N = 3412

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: pathlibr7iz.4xlarge48121620SE +/- 0.03, N = 314.5

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: nbodyr7iz.4xlarge20406080100SE +/- 0.12, N = 396.2

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: floatr7iz.4xlarge20406080100SE +/- 0.03, N = 387.5

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: chaosr7iz.4xlarge20406080100SE +/- 0.09, N = 391.5

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: 2to3r7iz.4xlarge60120180240300265

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: gor7iz.4xlarge50100150200250208

x265

This is a simple test of the x265 encoder run on the CPU with 1080p and 4K options for H.265 video encode performance with x265. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betterx265 3.4Video Input: Bosphorus 1080pr7iz.4xlarge1530456075SE +/- 0.10, N = 368.091. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl

OpenBenchmarking.orgFrames Per Second, More Is Betterx265 3.4Video Input: Bosphorus 4Kr7iz.4xlarge48121620SE +/- 0.15, N = 615.271. (CXX) g++ options: -O3 -rdynamic -lpthread -lrt -ldl

x264

This is a multi-threaded test of the x264 video encoder run on the CPU with a choice of 1080p or 4K video input. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is Betterx264 2022-02-22Video Input: Bosphorus 1080pr7iz.4xlarge306090120150SE +/- 0.16, N = 3127.151. (CC) gcc options: -ldl -m64 -lm -lpthread -O3 -flto

OpenBenchmarking.orgFrames Per Second, More Is Betterx264 2022-02-22Video Input: Bosphorus 4Kr7iz.4xlarge714212835SE +/- 0.07, N = 330.231. (CC) gcc options: -ldl -m64 -lm -lpthread -O3 -flto

Botan

Botan is a BSD-licensed cross-platform open-source C++ crypto library "cryptography toolkit" that supports most publicly known cryptographic algorithms. The project's stated goal is to be "the best option for cryptography in C++ by offering the tools necessary to implement a range of practical systems, such as TLS protocol, X.509 certificates, modern AEAD ciphers, PKCS#11 and TPM hardware support, password hashing, and post quantum crypto schemes." Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMiB/s, More Is BetterBotan 2.17.3Test: ChaCha20Poly1305 - Decryptr7iz.4xlarge2004006008001000SE +/- 0.20, N = 3846.951. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt

OpenBenchmarking.orgMiB/s, More Is BetterBotan 2.17.3Test: ChaCha20Poly1305r7iz.4xlarge2004006008001000SE +/- 0.95, N = 3845.751. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt

OpenBenchmarking.orgMiB/s, More Is BetterBotan 2.17.3Test: CAST-256 - Decryptr7iz.4xlarge306090120150SE +/- 0.00, N = 3139.371. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt

OpenBenchmarking.orgMiB/s, More Is BetterBotan 2.17.3Test: CAST-256r7iz.4xlarge306090120150SE +/- 0.00, N = 3139.271. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt

OpenBenchmarking.orgMiB/s, More Is BetterBotan 2.17.3Test: Blowfish - Decryptr7iz.4xlarge100200300400500SE +/- 0.04, N = 3477.761. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt

OpenBenchmarking.orgMiB/s, More Is BetterBotan 2.17.3Test: Blowfishr7iz.4xlarge100200300400500SE +/- 0.01, N = 3478.711. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt

OpenBenchmarking.orgMiB/s, More Is BetterBotan 2.17.3Test: Twofish - Decryptr7iz.4xlarge80160240320400SE +/- 0.86, N = 3387.681. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt

OpenBenchmarking.orgMiB/s, More Is BetterBotan 2.17.3Test: Twofishr7iz.4xlarge80160240320400SE +/- 0.71, N = 3387.261. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt

OpenBenchmarking.orgMiB/s, More Is BetterBotan 2.17.3Test: AES-256 - Decryptr7iz.4xlarge12002400360048006000SE +/- 2.77, N = 35505.441. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt

OpenBenchmarking.orgMiB/s, More Is BetterBotan 2.17.3Test: AES-256r7iz.4xlarge12002400360048006000SE +/- 3.00, N = 35529.151. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt

OpenBenchmarking.orgMiB/s, More Is BetterBotan 2.17.3Test: KASUMI - Decryptr7iz.4xlarge20406080100SE +/- 0.15, N = 391.061. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt

OpenBenchmarking.orgMiB/s, More Is BetterBotan 2.17.3Test: KASUMIr7iz.4xlarge20406080100SE +/- 0.01, N = 392.751. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt

OpenSSL

OpenSSL is an open-source toolkit that implements SSL (Secure Sockets Layer) and TLS (Transport Layer Security) protocols. This test profile makes use of the built-in "openssl speed" benchmarking capabilities. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgbyte/s, More Is BetterOpenSSL 3.1Algorithm: ChaCha20-Poly1305r7iz.4xlarge9000M18000M27000M36000M45000MSE +/- 53224818.96, N = 3415910394071. (CC) gcc options: -pthread -m64 -O3 -lssl -lcrypto -ldl

OpenBenchmarking.orgbyte/s, More Is BetterOpenSSL 3.1Algorithm: AES-256-GCMr7iz.4xlarge30000M60000M90000M120000M150000MSE +/- 338177292.11, N = 31275062312031. (CC) gcc options: -pthread -m64 -O3 -lssl -lcrypto -ldl

OpenBenchmarking.orgbyte/s, More Is BetterOpenSSL 3.1Algorithm: AES-128-GCMr7iz.4xlarge40000M80000M120000M160000M200000MSE +/- 192097685.01, N = 31663226232831. (CC) gcc options: -pthread -m64 -O3 -lssl -lcrypto -ldl

OpenBenchmarking.orgbyte/s, More Is BetterOpenSSL 3.1Algorithm: ChaCha20r7iz.4xlarge12000M24000M36000M48000M60000MSE +/- 23433776.10, N = 3578492210301. (CC) gcc options: -pthread -m64 -O3 -lssl -lcrypto -ldl

OpenBenchmarking.orgverify/s, More Is BetterOpenSSL 3.1Algorithm: RSA4096r7iz.4xlarge40K80K120K160K200KSE +/- 199.29, N = 3175992.61. (CC) gcc options: -pthread -m64 -O3 -lssl -lcrypto -ldl

OpenBenchmarking.orgsign/s, More Is BetterOpenSSL 3.1Algorithm: RSA4096r7iz.4xlarge13002600390052006500SE +/- 2.30, N = 35961.41. (CC) gcc options: -pthread -m64 -O3 -lssl -lcrypto -ldl

OpenBenchmarking.orgbyte/s, More Is BetterOpenSSL 3.1Algorithm: SHA512r7iz.4xlarge900M1800M2700M3600M4500MSE +/- 994026.65, N = 341379829731. (CC) gcc options: -pthread -m64 -O3 -lssl -lcrypto -ldl

OpenBenchmarking.orgbyte/s, More Is BetterOpenSSL 3.1Algorithm: SHA256r7iz.4xlarge3000M6000M9000M12000M15000MSE +/- 14860272.79, N = 3116779254671. (CC) gcc options: -pthread -m64 -O3 -lssl -lcrypto -ldl

LZ4 Compression

This test measures the time needed to compress/decompress a sample file (an Ubuntu ISO) using LZ4 compression. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 9 - Decompression Speedr7iz.4xlarge2K4K6K8K10KSE +/- 0.74, N = 38565.71. (CC) gcc options: -O3

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 9 - Compression Speedr7iz.4xlarge1326395265SE +/- 0.01, N = 357.131. (CC) gcc options: -O3

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 3 - Decompression Speedr7iz.4xlarge2K4K6K8K10KSE +/- 1.85, N = 38559.11. (CC) gcc options: -O3

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 3 - Compression Speedr7iz.4xlarge1326395265SE +/- 0.00, N = 358.451. (CC) gcc options: -O3

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 1 - Decompression Speedr7iz.4xlarge2K4K6K8K10KSE +/- 34.33, N = 38859.61. (CC) gcc options: -O3

OpenBenchmarking.orgMB/s, More Is BetterLZ4 Compression 1.9.3Compression Level: 1 - Compression Speedr7iz.4xlarge2K4K6K8K10KSE +/- 30.43, N = 38081.751. (CC) gcc options: -O3

Zstd Compression

This test measures the time needed to compress/decompress a sample file (silesia.tar) using Zstd (Zstandard) compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19, Long Mode - Decompression Speedr7iz.4xlarge30060090012001500SE +/- 1.51, N = 31306.21. (CC) gcc options: -O3 -pthread -lz

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19, Long Mode - Compression Speedr7iz.4xlarge246810SE +/- 0.01, N = 37.781. (CC) gcc options: -O3 -pthread -lz

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8, Long Mode - Decompression Speedr7iz.4xlarge30060090012001500SE +/- 1.45, N = 31619.51. (CC) gcc options: -O3 -pthread -lz

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8, Long Mode - Compression Speedr7iz.4xlarge130260390520650SE +/- 0.94, N = 3581.21. (CC) gcc options: -O3 -pthread -lz

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3, Long Mode - Decompression Speedr7iz.4xlarge30060090012001500SE +/- 0.53, N = 31577.61. (CC) gcc options: -O3 -pthread -lz

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3, Long Mode - Compression Speedr7iz.4xlarge2004006008001000SE +/- 1.83, N = 3903.21. (CC) gcc options: -O3 -pthread -lz

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19 - Decompression Speedr7iz.4xlarge30060090012001500SE +/- 2.75, N = 31357.51. (CC) gcc options: -O3 -pthread -lz

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 19 - Compression Speedr7iz.4xlarge48121620SE +/- 0.00, N = 314.41. (CC) gcc options: -O3 -pthread -lz

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 12 - Decompression Speedr7iz.4xlarge30060090012001500SE +/- 1.39, N = 31569.01. (CC) gcc options: -O3 -pthread -lz

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 12 - Compression Speedr7iz.4xlarge50100150200250SE +/- 2.00, N = 3219.11. (CC) gcc options: -O3 -pthread -lz

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8 - Decompression Speedr7iz.4xlarge400800120016002000SE +/- 0.62, N = 31637.51. (CC) gcc options: -O3 -pthread -lz

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 8 - Compression Speedr7iz.4xlarge120240360480600SE +/- 5.39, N = 3552.01. (CC) gcc options: -O3 -pthread -lz

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3 - Decompression Speedr7iz.4xlarge30060090012001500SE +/- 0.38, N = 31538.81. (CC) gcc options: -O3 -pthread -lz

OpenBenchmarking.orgMB/s, More Is BetterZstd Compression 1.5.4Compression Level: 3 - Compression Speedr7iz.4xlarge5001000150020002500SE +/- 13.48, N = 32305.81. (CC) gcc options: -O3 -pthread -lz

GNU GMP GMPbench

GMPbench is a test of the GNU Multiple Precision Arithmetic (GMP) Library. GMPbench is a single-threaded integer benchmark that leverages the GMP library to stress the CPU with widening integer multiplication. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGMPbench Score, More Is BetterGNU GMP GMPbench 6.2.1Total Timer7iz.4xlarge140028004200560070006370.71. (CC) gcc options: -O3 -fomit-frame-pointer -lm

CacheBench

This is a performance test of CacheBench, which is part of LLCbench. CacheBench is designed to test the memory and cache bandwidth performance Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterCacheBenchTest: Read / Modify / Writer7iz.4xlarge12K24K36K48K60KSE +/- 153.63, N = 356544.65MIN: 49841.95 / MAX: 58959.951. (CC) gcc options: -lrt

OpenBenchmarking.orgMB/s, More Is BetterCacheBenchTest: Writer7iz.4xlarge6K12K18K24K30KSE +/- 3.34, N = 328547.30MIN: 25912.59 / MAX: 29653.951. (CC) gcc options: -lrt

OpenBenchmarking.orgMB/s, More Is BetterCacheBenchTest: Readr7iz.4xlarge9001800270036004500SE +/- 0.02, N = 34246.07MIN: 4244.89 / MAX: 4246.231. (CC) gcc options: -lrt

Intel Memory Latency Checker

Intel Memory Latency Checker (MLC) is a binary-only system memory bandwidth and latency benchmark. If the download fails you may need to manually download the file from https://www.intel.com/content/www/us/en/developer/articles/tool/intelr-memory-latency-checker.html and place it in your PTS download cache. On some systems root privileges are needed to run the MLC tester. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterIntel Memory Latency Checker 3.10Test: Peak Injection Bandwidth - Stream-Triad Liker7iz.4xlarge40K80K120K160K200KSE +/- 942.58, N = 3167090.0

OpenBenchmarking.orgMB/s, More Is BetterIntel Memory Latency Checker 3.10Test: Peak Injection Bandwidth - 1:1 Reads-Writesr7iz.4xlarge40K80K120K160K200KSE +/- 2056.94, N = 3173449.5

OpenBenchmarking.orgMB/s, More Is BetterIntel Memory Latency Checker 3.10Test: Peak Injection Bandwidth - 2:1 Reads-Writesr7iz.4xlarge40K80K120K160K200KSE +/- 1491.77, N = 3174952.4

OpenBenchmarking.orgMB/s, More Is BetterIntel Memory Latency Checker 3.10Test: Peak Injection Bandwidth - 3:1 Reads-Writesr7iz.4xlarge40K80K120K160K200KSE +/- 1462.57, N = 3169108.2

OpenBenchmarking.orgMB/s, More Is BetterIntel Memory Latency Checker 3.10Test: Peak Injection Bandwidth - All Readsr7iz.4xlarge30K60K90K120K150KSE +/- 1102.17, N = 3147899.5

OpenBenchmarking.orgMB/s, More Is BetterIntel Memory Latency Checker 3.10Test: Max Bandwidth - Stream-Triad Liker7iz.4xlarge40K80K120K160K200KSE +/- 465.38, N = 3167806.81

OpenBenchmarking.orgMB/s, More Is BetterIntel Memory Latency Checker 3.10Test: Max Bandwidth - 1:1 Reads-Writesr7iz.4xlarge40K80K120K160K200KSE +/- 140.92, N = 3176001.88

OpenBenchmarking.orgMB/s, More Is BetterIntel Memory Latency Checker 3.10Test: Max Bandwidth - 2:1 Reads-Writesr7iz.4xlarge40K80K120K160K200KSE +/- 329.83, N = 3176424.12

OpenBenchmarking.orgMB/s, More Is BetterIntel Memory Latency Checker 3.10Test: Max Bandwidth - 3:1 Reads-Writesr7iz.4xlarge40K80K120K160K200KSE +/- 158.09, N = 3170862.56

OpenBenchmarking.orgMB/s, More Is BetterIntel Memory Latency Checker 3.10Test: Max Bandwidth - All Readsr7iz.4xlarge30K60K90K120K150KSE +/- 76.86, N = 3149243.59

OpenBenchmarking.orgns, Fewer Is BetterIntel Memory Latency Checker 3.10Test: Idle Latencyr7iz.4xlarge306090120150SE +/- 0.03, N = 3113.9

Stream

This is a benchmark of Stream, the popular system memory (RAM) benchmark. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMB/s, More Is BetterStream 2013-01-17Type: Addr7iz.4xlarge30K60K90K120K150KSE +/- 15.19, N = 5128078.91. (CC) gcc options: -mcmodel=medium -O3 -march=native -fopenmp

OpenBenchmarking.orgMB/s, More Is BetterStream 2013-01-17Type: Triadr7iz.4xlarge30K60K90K120K150KSE +/- 5.01, N = 5128231.51. (CC) gcc options: -mcmodel=medium -O3 -march=native -fopenmp

OpenBenchmarking.orgMB/s, More Is BetterStream 2013-01-17Type: Scaler7iz.4xlarge30K60K90K120K150KSE +/- 21.09, N = 5118927.71. (CC) gcc options: -mcmodel=medium -O3 -march=native -fopenmp

OpenBenchmarking.orgMB/s, More Is BetterStream 2013-01-17Type: Copyr7iz.4xlarge40K80K120K160K200KSE +/- 29.80, N = 5171559.61. (CC) gcc options: -mcmodel=medium -O3 -march=native -fopenmp

Apache Spark

This is a benchmark of Apache Spark with its PySpark interface. Apache Spark is an open-source unified analytics engine for large-scale data processing and dealing with big data. This test profile benchmars the Apache Spark in a single-system configuration using spark-submit. The test makes use of DIYBigData's pyspark-benchmark (https://github.com/DIYBigData/pyspark-benchmark/) for generating of test data and various Apache Spark operations. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 2000 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge3691215SE +/- 0.68, N = 312.36

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 40000000 - Partitions: 500 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge48121620SE +/- 1.17, N = 314.27

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 20000000 - Partitions: 500 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge48121620SE +/- 1.17, N = 315.12

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 10000000 - Partitions: 500 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge48121620SE +/- 0.62, N = 314.33

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 500 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge48121620SE +/- 1.45, N = 314.51

OpenBenchmarking.orgSeconds, Fewer Is BetterApache Spark 3.3Row Count: 1000000 - Partitions: 100 - Calculate Pi Benchmark Using Dataframer7iz.4xlarge3691215SE +/- 0.48, N = 312.35

NCNN

NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: Vulkan GPU - Model: resnet18r7iz.4xlarge0.96981.93962.90943.87924.849SE +/- 0.19, N = 34.31MIN: 3.8 / MAX: 9.331. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: squeezenet_ssdr7iz.4xlarge246810SE +/- 0.13, N = 86.04MIN: 5.55 / MAX: 10.951. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: yolov4-tinyr7iz.4xlarge48121620SE +/- 0.38, N = 814.17MIN: 13.08 / MAX: 36.591. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: resnet18r7iz.4xlarge0.93151.8632.79453.7264.6575SE +/- 0.12, N = 84.14MIN: 3.81 / MAX: 9.731. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

OpenBenchmarking.orgms, Fewer Is BetterNCNN 20230517Target: CPU - Model: googlenetr7iz.4xlarge246810SE +/- 0.14, N = 86.39MIN: 5.95 / MAX: 20.651. (CXX) g++ options: -O3 -rdynamic -lgomp -lpthread -pthread

nginx

This is a benchmark of the lightweight Nginx HTTP(S) web-server. This Nginx web server benchmark test profile makes use of the wrk program for facilitating the HTTP requests over a fixed period time with a configurable number of concurrent clients/connections. HTTPS with a self-signed OpenSSL certificate is used by this test for local benchmarking. Learn more via the OpenBenchmarking.org test page.

Connections: 1

r7iz.4xlarge: The test quit with a non-zero exit status.

DaCapo Benchmark

This test runs the DaCapo Benchmarks written in Java and intended to test system/CPU performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgmsec, Fewer Is BetterDaCapo Benchmark 9.12-MR1Java Test: H2r7iz.4xlarge7001400210028003500SE +/- 136.02, N = 163255

Renaissance

Renaissance is a suite of benchmarks designed to test the Java JVM from Apache Spark to a Twitter-like service to Scala and other features. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterRenaissance 0.14Test: Savina Reactors.IOr7iz.4xlarge16003200480064008000SE +/- 113.44, N = 157237.0MIN: 6584.57 / MAX: 15031.61

OpenBenchmarking.orgms, Fewer Is BetterRenaissance 0.14Test: Apache Spark ALSr7iz.4xlarge5001000150020002500SE +/- 60.58, N = 152531.3MIN: 1981.85 / MAX: 2944.31

296 Results Shown

Apache Spark:
  40000000 - 2000 - Broadcast Inner Join Test Time
  40000000 - 2000 - Inner Join Test Time
  40000000 - 2000 - Repartition Test Time
  40000000 - 2000 - Group By Test Time
  40000000 - 2000 - Calculate Pi Benchmark Using Dataframe
  40000000 - 2000 - Calculate Pi Benchmark
  40000000 - 2000 - SHA-512 Benchmark Time
  40000000 - 1000 - Broadcast Inner Join Test Time
  40000000 - 1000 - Inner Join Test Time
  40000000 - 1000 - Repartition Test Time
  40000000 - 1000 - Group By Test Time
  40000000 - 1000 - Calculate Pi Benchmark Using Dataframe
  40000000 - 1000 - Calculate Pi Benchmark
  40000000 - 1000 - SHA-512 Benchmark Time
  20000000 - 2000 - Broadcast Inner Join Test Time
  20000000 - 2000 - Inner Join Test Time
  20000000 - 2000 - Repartition Test Time
  20000000 - 2000 - Group By Test Time
  20000000 - 2000 - Calculate Pi Benchmark
  20000000 - 2000 - SHA-512 Benchmark Time
  20000000 - 1000 - Broadcast Inner Join Test Time
  20000000 - 1000 - Inner Join Test Time
  20000000 - 1000 - Repartition Test Time
  20000000 - 1000 - Group By Test Time
  20000000 - 1000 - Calculate Pi Benchmark Using Dataframe
  20000000 - 1000 - Calculate Pi Benchmark
  20000000 - 1000 - SHA-512 Benchmark Time
  10000000 - 2000 - Broadcast Inner Join Test Time
  10000000 - 2000 - Inner Join Test Time
  10000000 - 2000 - Repartition Test Time
  10000000 - 2000 - Group By Test Time
  10000000 - 2000 - Calculate Pi Benchmark Using Dataframe
  10000000 - 2000 - Calculate Pi Benchmark
  10000000 - 2000 - SHA-512 Benchmark Time
  10000000 - 1000 - Broadcast Inner Join Test Time
  10000000 - 1000 - Inner Join Test Time
  10000000 - 1000 - Repartition Test Time
  10000000 - 1000 - Group By Test Time
  10000000 - 1000 - Calculate Pi Benchmark Using Dataframe
  10000000 - 1000 - Calculate Pi Benchmark
  10000000 - 1000 - SHA-512 Benchmark Time
  40000000 - 500 - Broadcast Inner Join Test Time
  40000000 - 500 - Inner Join Test Time
  40000000 - 500 - Repartition Test Time
  40000000 - 500 - Group By Test Time
  40000000 - 500 - Calculate Pi Benchmark
  40000000 - 500 - SHA-512 Benchmark Time
  40000000 - 100 - Broadcast Inner Join Test Time
  40000000 - 100 - Inner Join Test Time
  40000000 - 100 - Repartition Test Time
  40000000 - 100 - Group By Test Time
  40000000 - 100 - Calculate Pi Benchmark Using Dataframe
  40000000 - 100 - Calculate Pi Benchmark
  40000000 - 100 - SHA-512 Benchmark Time
  20000000 - 500 - Broadcast Inner Join Test Time
  20000000 - 500 - Inner Join Test Time
  20000000 - 500 - Repartition Test Time
  20000000 - 500 - Group By Test Time
  20000000 - 500 - Calculate Pi Benchmark
  20000000 - 500 - SHA-512 Benchmark Time
  20000000 - 100 - Broadcast Inner Join Test Time
  20000000 - 100 - Inner Join Test Time
  20000000 - 100 - Repartition Test Time
  20000000 - 100 - Group By Test Time
  20000000 - 100 - Calculate Pi Benchmark Using Dataframe
  20000000 - 100 - Calculate Pi Benchmark
  20000000 - 100 - SHA-512 Benchmark Time
  10000000 - 500 - Broadcast Inner Join Test Time
  10000000 - 500 - Inner Join Test Time
  10000000 - 500 - Repartition Test Time
  10000000 - 500 - Group By Test Time
  10000000 - 500 - Calculate Pi Benchmark
  10000000 - 500 - SHA-512 Benchmark Time
  10000000 - 100 - Broadcast Inner Join Test Time
  10000000 - 100 - Inner Join Test Time
  10000000 - 100 - Repartition Test Time
  10000000 - 100 - Group By Test Time
  10000000 - 100 - Calculate Pi Benchmark Using Dataframe
  10000000 - 100 - Calculate Pi Benchmark
  10000000 - 100 - SHA-512 Benchmark Time
  1000000 - 2000 - Broadcast Inner Join Test Time
  1000000 - 2000 - Inner Join Test Time
  1000000 - 2000 - Repartition Test Time
  1000000 - 2000 - Group By Test Time
  1000000 - 2000 - Calculate Pi Benchmark Using Dataframe
  1000000 - 2000 - Calculate Pi Benchmark
  1000000 - 2000 - SHA-512 Benchmark Time
  1000000 - 1000 - Broadcast Inner Join Test Time
  1000000 - 1000 - Inner Join Test Time
  1000000 - 1000 - Repartition Test Time
  1000000 - 1000 - Group By Test Time
  1000000 - 1000 - Calculate Pi Benchmark Using Dataframe
  1000000 - 1000 - Calculate Pi Benchmark
  1000000 - 1000 - SHA-512 Benchmark Time
  1000000 - 500 - Broadcast Inner Join Test Time
  1000000 - 500 - Inner Join Test Time
  1000000 - 500 - Repartition Test Time
  1000000 - 500 - Group By Test Time
  1000000 - 500 - Calculate Pi Benchmark
  1000000 - 500 - SHA-512 Benchmark Time
  1000000 - 100 - Broadcast Inner Join Test Time
  1000000 - 100 - Inner Join Test Time
  1000000 - 100 - Repartition Test Time
  1000000 - 100 - Group By Test Time
  1000000 - 100 - Calculate Pi Benchmark
  1000000 - 100 - SHA-512 Benchmark Time
libxsmm:
  64
  32
  256
  128
NCNN:
  Vulkan GPU - alexnet
  Vulkan GPU - FastestDet
  Vulkan GPU - vision_transformer
  Vulkan GPU - regnety_400m
  Vulkan GPU - squeezenet_ssd
  Vulkan GPU - yolov4-tiny
  Vulkan GPU - resnet50
  Vulkan GPU - vgg16
  Vulkan GPU - googlenet
  Vulkan GPU - blazeface
  Vulkan GPU - efficientnet-b0
  Vulkan GPU - mnasnet
  Vulkan GPU - shufflenet-v2
  Vulkan GPU-v3-v3 - mobilenet-v3
  Vulkan GPU-v2-v2 - mobilenet-v2
  Vulkan GPU - mobilenet
  CPU - mnasnet
  CPU - FastestDet
  CPU - vision_transformer
  CPU - regnety_400m
  CPU - resnet50
  CPU - alexnet
  CPU - vgg16
  CPU - blazeface
  CPU - efficientnet-b0
  CPU - shufflenet-v2
  CPU-v3-v3 - mobilenet-v3
  CPU-v2-v2 - mobilenet-v2
  CPU - mobilenet
VVenC:
  Bosphorus 1080p - Faster
  Bosphorus 1080p - Fast
  Bosphorus 4K - Faster
  Bosphorus 4K - Fast
libjpeg-turbo tjbench
Apache Cassandra:
  Mixed 1:3
  Mixed 1:1
  Writes
Redis 7.0.12 + memtier_benchmark:
  Redis - 500 - 1:10
  Redis - 500 - 10:1
  Redis - 100 - 1:10
  Redis - 100 - 10:1
  Redis - 500 - 5:1
  Redis - 500 - 1:5
  Redis - 500 - 1:1
  Redis - 50 - 1:10
  Redis - 50 - 10:1
  Redis - 100 - 5:1
  Redis - 100 - 1:5
  Redis - 100 - 1:1
  Redis - 50 - 5:1
  Redis - 50 - 1:5
  Redis - 50 - 1:1
nginx:
  4000
  1000
  500
  200
  100
  20
BlogBench:
  Write
  Read
DaCapo Benchmark:
  Tradebeans
  Tradesoap
  Eclipse
  Jython
Renaissance:
  Genetic Algorithm Using Jenetics + Futures
  Akka Unbalanced Cobwebbed Tree
  In-Memory Database Shootout
  Finagle HTTP Requests
  Apache Spark PageRank
  Apache Spark Bayes
  ALS Movie Lens
  Rand Forest
  Scala Dotty
SciMark:
  Jacobi Successive Over-Relaxation
  Dense LU Matrix Factorization
  Sparse Matrix Multiply
  Fast Fourier Transform
  Monte Carlo
  Composite
C-Ray
Stockfish
Smallpt
GraphicsMagick:
  HWB Color Space
  Noise-Gaussian
  Resizing
  Enhanced
  Sharpen
  Rotate
  Swirl
CppPerformanceBenchmarks:
  Stepanov Abstraction
  Function Objects
  Stepanov Vector
  Rand Numbers
  Math Library
  Ctype
  Atol
PyPerformance:
  pickle_pure_python
  django_template
  python_startup
  regex_compile
  crypto_pyaes
  json_loads
  raytrace
  pathlib
  nbody
  float
  chaos
  2to3
  go
x265:
  Bosphorus 1080p
  Bosphorus 4K
x264:
  Bosphorus 1080p
  Bosphorus 4K
Botan:
  ChaCha20Poly1305 - Decrypt
  ChaCha20Poly1305
  CAST-256 - Decrypt
  CAST-256
  Blowfish - Decrypt
  Blowfish
  Twofish - Decrypt
  Twofish
  AES-256 - Decrypt
  AES-256
  KASUMI - Decrypt
  KASUMI
OpenSSL:
  ChaCha20-Poly1305
  AES-256-GCM
  AES-128-GCM
  ChaCha20
  RSA4096
  RSA4096
  SHA512
  SHA256
LZ4 Compression:
  9 - Decompression Speed
  9 - Compression Speed
  3 - Decompression Speed
  3 - Compression Speed
  1 - Decompression Speed
  1 - Compression Speed
Zstd Compression:
  19, Long Mode - Decompression Speed
  19, Long Mode - Compression Speed
  8, Long Mode - Decompression Speed
  8, Long Mode - Compression Speed
  3, Long Mode - Decompression Speed
  3, Long Mode - Compression Speed
  19 - Decompression Speed
  19 - Compression Speed
  12 - Decompression Speed
  12 - Compression Speed
  8 - Decompression Speed
  8 - Compression Speed
  3 - Decompression Speed
  3 - Compression Speed
GNU GMP GMPbench
CacheBench:
  Read / Modify / Write
  Write
  Read
Intel Memory Latency Checker:
  Peak Injection Bandwidth - Stream-Triad Like
  Peak Injection Bandwidth - 1:1 Reads-Writes
  Peak Injection Bandwidth - 2:1 Reads-Writes
  Peak Injection Bandwidth - 3:1 Reads-Writes
  Peak Injection Bandwidth - All Reads
  Max Bandwidth - Stream-Triad Like
  Max Bandwidth - 1:1 Reads-Writes
  Max Bandwidth - 2:1 Reads-Writes
  Max Bandwidth - 3:1 Reads-Writes
  Max Bandwidth - All Reads
  Idle Latency
Stream:
  Add
  Triad
  Scale
  Copy
Apache Spark:
  20000000 - 2000 - Calculate Pi Benchmark Using Dataframe
  40000000 - 500 - Calculate Pi Benchmark Using Dataframe
  20000000 - 500 - Calculate Pi Benchmark Using Dataframe
  10000000 - 500 - Calculate Pi Benchmark Using Dataframe
  1000000 - 500 - Calculate Pi Benchmark Using Dataframe
  1000000 - 100 - Calculate Pi Benchmark Using Dataframe
NCNN:
  Vulkan GPU - resnet18
  CPU - squeezenet_ssd
  CPU - yolov4-tiny
  CPU - resnet18
  CPU - googlenet
DaCapo Benchmark
Renaissance:
  Savina Reactors.IO
  Apache Spark ALS