2020-02-29_-_btrd3_-_disk

Intel Core i5-3470 testing with a Gigabyte B75M-D3V (F11 BIOS) and NVIDIA GeForce GTX 760 2GB on Ubuntu 19.10 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2002292-VE-20200229B02
Jump To Table - Results

Statistics

Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
2020-02-29_-_btrd3_-_disk
February 29 2020
  7 Hours, 29 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


2020-02-29_-_btrd3_-_diskOpenBenchmarking.orgPhoronix Test SuiteIntel Core i5-3470 @ 3.60GHz (4 Cores)Gigabyte B75M-D3V (F11 BIOS)Intel Xeon E3-1200 v2/3rd8GB1000GB TOSHIBA MQ01ABD1 + 256GB OCZ PETROL + 750GB SAMSUNG HD753LJNVIDIA GeForce GTX 760 2GB (980/3004MHz)Realtek ALC887-VDRealtek RTL8111/8168/8411Ubuntu 19.105.3.0-40-generic (x86_64)GNOME Shell 3.34.1X Server 1.20.5NVIDIA 435.214.6.0GCC 9.2.1 20191008ext41920x1080ProcessorMotherboardChipsetMemoryDiskGraphicsAudioNetworkOSKernelDesktopDisplay ServerDisplay DriverOpenGLCompilerFile-SystemScreen Resolution2020-02-29_-_btrd3_-_disk BenchmarksSystem Logs- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-offload-targets=nvptx-none,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: intel_pstate powersave - CPU Microcode: 0x21- GPU Compute Cores: 1152- Python 2.7.17 + Python 3.7.5- itlb_multihit: KVM: Mitigation of Split huge pages + l1tf: Mitigation of PTE Inversion; VMX: conditional cache flushes SMT disabled + mds: Mitigation of Clear buffers; SMT disabled + meltdown: Mitigation of PTI + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full generic retpoline IBPB: conditional IBRS_FW STIBP: disabled RSB filling + tsx_async_abort: Not affected

2020-02-29_-_btrd3_-_diskmkl-dnn: IP Batch 1D - f32mkl-dnn: IP Batch All - f32mkl-dnn: IP Batch 1D - u8s8f32mkl-dnn: IP Batch All - u8s8f32mkl-dnn: Convolution Batch conv_3d - f32mkl-dnn: Convolution Batch conv_all - f32mkl-dnn: Convolution Batch conv_3d - u8s8f32mkl-dnn: Deconvolution Batch deconv_1d - f32mkl-dnn: Deconvolution Batch deconv_3d - f32mkl-dnn: Convolution Batch conv_alexnet - f32mkl-dnn: Convolution Batch conv_all - u8s8f32mkl-dnn: Deconvolution Batch deconv_all - f32mkl-dnn: Deconvolution Batch deconv_1d - u8s8f32mkl-dnn: Deconvolution Batch deconv_3d - u8s8f32mkl-dnn: Recurrent Neural Network Training - f32mkl-dnn: Convolution Batch conv_alexnet - u8s8f32mkl-dnn: Convolution Batch conv_googlenet_v3 - f32mkl-dnn: Convolution Batch conv_googlenet_v3 - u8s8f32numpy: rbenchmark: plaidml: No - Inference - VGG16 - CPUplaidml: No - Inference - ResNet 50 - CPUmlpack: scikit_icamlpack: scikit_svmscikit-learn: 2020-02-29_-_btrd3_-_disk50.716574.6921257.6711278.9780.047413889.546209.742.229560.93041939.2442050613690.718668.729153.21855.9914656.0784.3359775.98205.500.27512.623.4968.2515.0715.439OpenBenchmarking.org

MKL-DNN DNNL

This is a test of the Intel MKL-DNN (DNNL / Deep Neural Network Library) as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch 1D - Data Type: f322020-02-29_-_btrd3_-_disk1122334455SE +/- 0.29, N = 350.72MIN: 47.31. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch All - Data Type: f322020-02-29_-_btrd3_-_disk20406080100SE +/- 0.22, N = 374.69MIN: 69.861. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch 1D - Data Type: u8s8f322020-02-29_-_btrd3_-_disk60120180240300SE +/- 0.20, N = 3257.67MIN: 249.111. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: IP Batch All - Data Type: u8s8f322020-02-29_-_btrd3_-_disk30060090012001500SE +/- 0.28, N = 31278.97MIN: 1250.861. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_3d - Data Type: f322020-02-29_-_btrd3_-_disk20406080100SE +/- 0.11, N = 380.05MIN: 74.151. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_all - Data Type: f322020-02-29_-_btrd3_-_disk3K6K9K12K15KSE +/- 15.67, N = 313889.5MIN: 13588.71. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_3d - Data Type: u8s8f322020-02-29_-_btrd3_-_disk10K20K30K40K50KSE +/- 25.61, N = 346209.7MIN: 45833.11. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_1d - Data Type: f322020-02-29_-_btrd3_-_disk1020304050SE +/- 0.07, N = 342.23MIN: 39.471. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_3d - Data Type: f322020-02-29_-_btrd3_-_disk1428425670SE +/- 0.60, N = 360.93MIN: 57.271. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_alexnet - Data Type: f322020-02-29_-_btrd3_-_disk400800120016002000SE +/- 0.95, N = 31939.24MIN: 1911.651. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_all - Data Type: u8s8f322020-02-29_-_btrd3_-_disk90K180K270K360K450KSE +/- 2544.72, N = 3420506MIN: 4118001. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_all - Data Type: f322020-02-29_-_btrd3_-_disk3K6K9K12K15KSE +/- 16.26, N = 313690.7MIN: 13331.41. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_1d - Data Type: u8s8f322020-02-29_-_btrd3_-_disk4K8K12K16K20KSE +/- 53.74, N = 318668.7MIN: 18449.11. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Deconvolution Batch deconv_3d - Data Type: u8s8f322020-02-29_-_btrd3_-_disk6K12K18K24K30KSE +/- 31.35, N = 329153.2MIN: 28833.71. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Recurrent Neural Network Training - Data Type: f322020-02-29_-_btrd3_-_disk400800120016002000SE +/- 2.87, N = 31855.99MIN: 1809.591. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_alexnet - Data Type: u8s8f322020-02-29_-_btrd3_-_disk3K6K9K12K15KSE +/- 25.14, N = 314656.0MIN: 14521.61. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_googlenet_v3 - Data Type: f322020-02-29_-_btrd3_-_disk2004006008001000SE +/- 0.55, N = 3784.34MIN: 743.531. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

OpenBenchmarking.orgms, Fewer Is BetterMKL-DNN DNNL 1.1Harness: Convolution Batch conv_googlenet_v3 - Data Type: u8s8f322020-02-29_-_btrd3_-_disk2K4K6K8K10KSE +/- 19.19, N = 39775.98MIN: 9587.971. (CXX) g++ options: -O3 -march=native -std=c++11 -msse4.1 -fPIC -fopenmp -pie -lpthread -lrt -ldl

Numpy Benchmark

This is a test to obtain the general Numpy performance. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgScore, More Is BetterNumpy Benchmark2020-02-29_-_btrd3_-_disk50100150200250SE +/- 0.18, N = 3205.50

R Benchmark

This test is a quick-running survey of general R performance Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterR Benchmark2020-02-29_-_btrd3_-_disk0.06190.12380.18570.24760.3095SE +/- 0.0018, N = 30.27511. R scripting front-end version 3.6.1 (2019-07-05)

PlaidML

This test profile uses PlaidML deep learning framework developed by Intel for offering up various benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: VGG16 - Device: CPU2020-02-29_-_btrd3_-_disk0.58951.1791.76852.3582.9475SE +/- 0.04, N = 92.62

OpenBenchmarking.orgFPS, More Is BetterPlaidMLFP16: No - Mode: Inference - Network: ResNet 50 - Device: CPU2020-02-29_-_btrd3_-_disk0.78531.57062.35593.14123.9265SE +/- 0.01, N = 33.49

Mlpack Benchmark

Mlpack benchmark scripts for machine learning libraries Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterMlpack BenchmarkBenchmark: scikit_ica2020-02-29_-_btrd3_-_disk1530456075SE +/- 0.11, N = 368.25

OpenBenchmarking.orgSeconds, Fewer Is BetterMlpack BenchmarkBenchmark: scikit_svm2020-02-29_-_btrd3_-_disk48121620SE +/- 0.01, N = 315.07

Scikit-Learn

Scikit-learn is a Python module for machine learning Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 0.22.12020-02-29_-_btrd3_-_disk48121620SE +/- 0.26, N = 1515.44