ARMv7 rev 3 testing with a ODROID-XU4 Hardkernel Odroid XU4 and llvmpipe 2GB on Ubuntu 18.04 via the Phoronix Test Suite.
Compare your own system(s) to this result file with the
Phoronix Test Suite by running the command:
phoronix-test-suite benchmark 1903179-HV-1903178SP62 Jetson Prep - Phoronix Test Suite Jetson Prep ARMv7 rev 3 testing with a ODROID-XU4 Hardkernel Odroid XU4 and llvmpipe 2GB on Ubuntu 18.04 via the Phoronix Test Suite.
HTML result view exported from: https://openbenchmarking.org/result/1903179-HV-1903178SP62&sro&grr .
Jetson Prep Processor Motherboard Memory Disk Graphics Monitor OS Kernel Desktop Display Server Display Driver OpenGL Vulkan Compiler File-System Screen Resolution Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q Raspberry Pi 3 Model B+ ASUS TinkerBoard Jetson TX1 Max-P ODROID-XU4 ARMv8 rev 0 @ 2.27GHz (8 Cores) jetson-xavier 16384MB 31GB HBG4a2 NVIDIA Tegra Xavier VE228 Ubuntu 18.04 4.9.108-tegra (aarch64) Unity 7.5.0 X Server 1.19.6 NVIDIA 31.0.2 4.6.0 1.1.76 GCC 7.3.0 + CUDA 10.0 ext4 1920x1080 ARMv8 rev 3 @ 2.04GHz (4 Cores / 6 Threads) quill 8192MB 31GB 032G34 NVIDIA TEGRA Ubuntu 16.04 4.4.38-tegra (aarch64) Unity 7.4.0 X Server 1.18.4 NVIDIA 28.2.1 4.5.0 GCC 5.4.0 20160609 + CUDA 9.0 ARMv8 rev 3 @ 1.27GHz (4 Cores / 6 Threads) ARMv7 rev 4 @ 1.40GHz (4 Cores) BCM2835 Raspberry Pi 3 Model B Plus Rev 1.3 926MB 32GB GB2MW BCM2708 Raspbian 9.6 4.19.23-v7+ (armv7l) LXDE X Server 1.19.2 GCC 6.3.0 20170516 656x416 ARMv7 rev 1 @ 1.80GHz (4 Cores) Rockchip (Device Tree) 2048MB 32GB GB1QT Debian 9.0 4.4.16-00006-g4431f98-dirty (armv7l) X Server 1.18.4 1024x768 ARMv8 rev 1 @ 1.73GHz (4 Cores) jetson_tx1 4096MB 16GB 016G32 NVIDIA Tegra X1 VE228 Ubuntu 16.04 4.4.38-tegra (aarch64) Unity 7.4.5 NVIDIA 28.1.0 4.5.0 1.0.8 GCC 5.4.0 20160609 1920x1080 ARMv7 rev 3 @ 1.50GHz (8 Cores) ODROID-XU4 Hardkernel Odroid XU4 2048MB 16GB AJTD4R llvmpipe 2GB Ubuntu 18.04 4.14.37-135 (armv7l) X Server 1.19.6 3.3 Mesa 18.0.0-rc5 (LLVM 6.0 128 bits) GCC 7.3.0 OpenBenchmarking.org Compiler Details - Jetson AGX Xavier: --build=aarch64-linux-gnu --disable-libquadmath --disable-libquadmath-support --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-fix-cortex-a53-843419 --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++ --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-nls --enable-plugin --enable-shared --enable-threads=posix --host=aarch64-linux-gnu --program-prefix=aarch64-linux-gnu- --target=aarch64-linux-gnu --with-default-libstdcxx-abi=new --with-gcc-major-version-only -v - Jetson TX2 Max-P: --build=aarch64-linux-gnu --disable-browser-plugin --disable-libquadmath --disable-werror --enable-checking=release --enable-clocale=gnu --enable-fix-cortex-a53-843419 --enable-gnu-unique-object --enable-gtk-cairo --enable-java-awt=gtk --enable-java-home --enable-languages=c,ada,c++,java,go,d,fortran,objc,obj-c++ --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-nls --enable-plugin --enable-shared --enable-threads=posix --host=aarch64-linux-gnu --target=aarch64-linux-gnu --with-arch-directory=aarch64 --with-default-libstdcxx-abi=new -v - Jetson TX2 Max-Q: --build=aarch64-linux-gnu --disable-browser-plugin --disable-libquadmath --disable-werror --enable-checking=release --enable-clocale=gnu --enable-fix-cortex-a53-843419 --enable-gnu-unique-object --enable-gtk-cairo --enable-java-awt=gtk --enable-java-home --enable-languages=c,ada,c++,java,go,d,fortran,objc,obj-c++ --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-nls --enable-plugin --enable-shared --enable-threads=posix --host=aarch64-linux-gnu --target=aarch64-linux-gnu --with-arch-directory=aarch64 --with-default-libstdcxx-abi=new -v - Raspberry Pi 3 Model B+: --build=arm-linux-gnueabihf --disable-browser-plugin --disable-libitm --disable-libquadmath --disable-sjlj-exceptions --enable-checking=release --enable-clocale=gnu --enable-gnu-unique-object --enable-gtk-cairo --enable-java-awt=gtk --enable-java-home --enable-languages=c,ada,c++,java,go,d,fortran,objc,obj-c++ --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-nls --enable-objc-gc=auto --enable-plugin --enable-shared --enable-threads=posix --host=arm-linux-gnueabihf --program-prefix=arm-linux-gnueabihf- --target=arm-linux-gnueabihf --with-arch-directory=arm --with-arch=armv6 --with-default-libstdcxx-abi=new --with-float=hard --with-fpu=vfp --with-target-system-zlib -v - ASUS TinkerBoard: --build=arm-linux-gnueabihf --disable-browser-plugin --disable-libitm --disable-libquadmath --disable-sjlj-exceptions --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-gtk-cairo --enable-java-awt=gtk --enable-java-home --enable-languages=c,ada,c++,java,go,d,fortran,objc,obj-c++ --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-nls --enable-objc-gc=auto --enable-plugin --enable-shared --enable-threads=posix --host=arm-linux-gnueabihf --program-prefix=arm-linux-gnueabihf- --target=arm-linux-gnueabihf --with-arch-directory=arm --with-arch=armv7-a --with-default-libstdcxx-abi=new --with-float=hard --with-fpu=vfpv3-d16 --with-mode=thumb --with-target-system-zlib -v - Jetson TX1 Max-P: --build=aarch64-linux-gnu --disable-browser-plugin --disable-libquadmath --disable-werror --enable-checking=release --enable-clocale=gnu --enable-fix-cortex-a53-843419 --enable-gnu-unique-object --enable-gtk-cairo --enable-java-awt=gtk --enable-java-home --enable-languages=c,ada,c++,java,go,d,fortran,objc,obj-c++ --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-nls --enable-plugin --enable-shared --enable-threads=posix --host=aarch64-linux-gnu --target=aarch64-linux-gnu --with-arch-directory=aarch64 --with-default-libstdcxx-abi=new -v - ODROID-XU4: --build=arm-linux-gnueabihf --disable-libitm --disable-libquadmath --disable-libquadmath-support --disable-sjlj-exceptions --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++ --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-multilib --enable-nls --enable-objc-gc=auto --enable-plugin --enable-shared --enable-threads=posix --host=arm-linux-gnueabihf --program-prefix=arm-linux-gnueabihf- --target=arm-linux-gnueabihf --with-arch=armv7-a --with-default-libstdcxx-abi=new --with-float=hard --with-fpu=vfpv3-d16 --with-gcc-major-version-only --with-mode=thumb --with-target-system-zlib -v Processor Details - Jetson AGX Xavier: Scaling Governor: tegra_cpufreq schedutil - Jetson TX2 Max-P: Scaling Governor: tegra_cpufreq schedutil - Jetson TX2 Max-Q: Scaling Governor: tegra_cpufreq schedutil - Raspberry Pi 3 Model B+: Scaling Governor: BCM2835 Freq ondemand - ASUS TinkerBoard: Scaling Governor: cpufreq-dt interactive - Jetson TX1 Max-P: Scaling Governor: tegra-cpufreq interactive - ODROID-XU4: Scaling Governor: cpufreq-dt ondemand Python Details - Jetson AGX Xavier: Python 2.7.15rc1 + Python 3.6.7 - Jetson TX2 Max-P: Python 2.7.12 + Python 3.5.2 - Jetson TX2 Max-Q: Python 2.7.12 + Python 3.5.2 - Raspberry Pi 3 Model B+: Python 2.7.13 + Python 3.5.3 - ASUS TinkerBoard: Python 2.7.13 + Python 3.5.3 - Jetson TX1 Max-P: Python 2.7.12 + Python 3.5.2 - ODROID-XU4: Python 2.7.15rc1 + Python 3.6.7 Kernel Details - ODROID-XU4: usbhid.quirks=0x0eef:0x0005:0x0004 Graphics Details - ODROID-XU4: EXA
Jetson Prep c-ray: Total Time - 4K, 16 Rays Per Pixel lczero: BLAS rust-prime: Prime Number Test To 200,000,000 ttsiod-renderer: Phong Rendering With Soft-Shadow Mapping cuda-mini-nbody: Original tensorrt-inference: ResNet152 - INT8 - 32 - Disabled tensorrt-inference: VGG19 - INT8 - 32 - Disabled pybench: Total For Average Test Times tensorrt-inference: VGG19 - FP16 - 32 - Disabled opencv-bench: tensorrt-inference: VGG16 - INT8 - 32 - Disabled tensorrt-inference: ResNet152 - FP16 - 32 - Disabled tensorrt-inference: VGG16 - FP16 - 32 - Disabled compress-zstd: Compressing ubuntu-16.04.3-server-i386.img, Compression Level 19 encode-flac: WAV To FLAC tensorrt-inference: ResNet50 - FP16 - 4 - Disabled tensorrt-inference: ResNet152 - INT8 - 4 - Disabled tensorrt-inference: VGG16 - INT8 - 4 - Disabled tensorrt-inference: ResNet152 - FP16 - 4 - Disabled compress-7zip: Compress Speed Test tensorrt-inference: VGG19 - INT8 - 4 - Disabled tensorrt-inference: ResNet50 - INT8 - 32 - Disabled glmark2: 1920 x 1080 tesseract-ocr: Time To OCR 7 Images tensorrt-inference: GoogleNet - FP16 - 32 - Disabled tensorrt-inference: VGG16 - FP16 - 4 - Disabled tensorrt-inference: VGG19 - FP16 - 4 - Disabled tensorrt-inference: GoogleNet - INT8 - 32 - Disabled tensorrt-inference: GoogleNet - FP16 - 4 - Disabled tensorrt-inference: ResNet50 - FP16 - 32 - Disabled tensorrt-inference: AlexNet - FP16 - 32 - Disabled tensorrt-inference: AlexNet - FP16 - 4 - Disabled tensorrt-inference: ResNet50 - INT8 - 4 - Disabled tensorrt-inference: GoogleNet - INT8 - 4 - Disabled tensorrt-inference: AlexNet - INT8 - 4 - Disabled tensorrt-inference: AlexNet - INT8 - 32 - Disabled lczero: CUDA + cuDNN lczero: CUDA + cuDNN FP16 Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q Raspberry Pi 3 Model B+ ASUS TinkerBoard Jetson TX1 Max-P ODROID-XU4 355 47.62 32.37 133 47.13 493.22 394.66 3007 203.96 128 475.08 259.82 247.95 80.06 54.47 547.50 372.73 303.78 224.19 19212 265.81 1215.08 2876 71.94 1006 208.76 172.50 1693 796 636 2038 1200 902.78 1146 1143 3143 952.89 2515.01 585 104.96 49.26 8.24 22.07 15.92 5408 29.83 296 19.91 41.91 36.87 144.97 65.07 92.28 18.29 17.56 35.11 5593 14.32 59.69 233 32.64 26.56 130 197 111 462 264 49.97 113 184 301 869 170.25 28.85 6.77 17.36 12.59 8735 23.94 493 15.79 32.67 29.83 253.80 104.28 72.01 14.50 14.24 27.34 3294 11.45 47.15 179 25.99 21.04 104 156 86.08 374 216 39.15 88.88 148 237 2030 1097.69 17.66 20913 2.74 342.23 339.53 2013 1718 1821.05 21.22 11502 496.62 279.05 2836 753 128.45 45.09 6339 145.80 79.20 4508 827 574.11 41.96 5009 520.70 97.03 4120 1 180.66 OpenBenchmarking.org
C-Ray Total Time - 4K, 16 Rays Per Pixel OpenBenchmarking.org Seconds, Fewer Is Better C-Ray 1.1 Total Time - 4K, 16 Rays Per Pixel ASUS TinkerBoard Jetson AGX Xavier Jetson TX1 Max-P Jetson TX2 Max-P Jetson TX2 Max-Q ODROID-XU4 Raspberry Pi 3 Model B+ 400 800 1200 1600 2000 SE +/- 22.09, N = 3 SE +/- 7.17, N = 9 SE +/- 10.23, N = 3 SE +/- 49.09, N = 9 SE +/- 1.44, N = 3 SE +/- 29.65, N = 9 SE +/- 2.46, N = 3 1718 355 753 585 869 827 2030 1. (CC) gcc options: -lm -lpthread -O3
LeelaChessZero Backend: BLAS OpenBenchmarking.org Nodes Per Second, More Is Better LeelaChessZero 0.20.1 Backend: BLAS Jetson AGX Xavier 11 22 33 44 55 SE +/- 0.62, N = 7 47.62 1. (CXX) g++ options: -lpthread -lz
Rust Prime Benchmark Prime Number Test To 200,000,000 OpenBenchmarking.org Seconds, Fewer Is Better Rust Prime Benchmark Prime Number Test To 200,000,000 ASUS TinkerBoard Jetson AGX Xavier Jetson TX1 Max-P Jetson TX2 Max-P Jetson TX2 Max-Q ODROID-XU4 Raspberry Pi 3 Model B+ 400 800 1200 1600 2000 SE +/- 187.90, N = 6 SE +/- 0.00, N = 3 SE +/- 0.77, N = 3 SE +/- 0.04, N = 3 SE +/- 0.09, N = 3 SE +/- 0.37, N = 3 SE +/- 1.55, N = 3 1821.05 32.37 128.45 104.96 170.25 574.11 1097.69 -ldl -lrt -lpthread -lgcc_s -lc -lm -lutil 1. (CC) gcc options: -pie -nodefaultlibs
TTSIOD 3D Renderer Phong Rendering With Soft-Shadow Mapping OpenBenchmarking.org FPS, More Is Better TTSIOD 3D Renderer 2.3b Phong Rendering With Soft-Shadow Mapping ASUS TinkerBoard Jetson AGX Xavier Jetson TX1 Max-P Jetson TX2 Max-P Jetson TX2 Max-Q ODROID-XU4 Raspberry Pi 3 Model B+ 30 60 90 120 150 SE +/- 0.27, N = 9 SE +/- 1.63, N = 12 SE +/- 0.04, N = 3 SE +/- 0.15, N = 3 SE +/- 0.46, N = 4 SE +/- 0.97, N = 9 SE +/- 0.16, N = 3 21.22 133.00 45.09 49.26 28.85 41.96 17.66 1. (CXX) g++ options: -O3 -fomit-frame-pointer -ffast-math -mtune=native -flto -lSDL -fopenmp -fwhole-program -lstdc++
CUDA Mini-Nbody Test: Original OpenBenchmarking.org (NBody^2)/s, More Is Better CUDA Mini-Nbody 2015-11-10 Test: Original Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 11 22 33 44 55 SE +/- 0.00, N = 3 SE +/- 0.01, N = 3 SE +/- 0.03, N = 3 47.13 8.24 6.77
NVIDIA TensorRT Inference Neural Network: ResNet152 - Precision: INT8 - Batch Size: 32 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: ResNet152 - Precision: INT8 - Batch Size: 32 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 110 220 330 440 550 SE +/- 0.81, N = 3 SE +/- 0.03, N = 3 SE +/- 0.00, N = 3 493.22 22.07 17.36
NVIDIA TensorRT Inference Neural Network: VGG19 - Precision: INT8 - Batch Size: 32 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: VGG19 - Precision: INT8 - Batch Size: 32 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 90 180 270 360 450 SE +/- 0.23, N = 3 SE +/- 0.06, N = 3 SE +/- 0.03, N = 3 394.66 15.92 12.59
PyBench Total For Average Test Times OpenBenchmarking.org Milliseconds, Fewer Is Better PyBench 2018-02-16 Total For Average Test Times ASUS TinkerBoard Jetson AGX Xavier Jetson TX1 Max-P Jetson TX2 Max-P Jetson TX2 Max-Q ODROID-XU4 Raspberry Pi 3 Model B+ 4K 8K 12K 16K 20K SE +/- 854.75, N = 9 SE +/- 4.67, N = 3 SE +/- 18.55, N = 3 SE +/- 33.86, N = 3 SE +/- 42.52, N = 3 SE +/- 30.99, N = 3 SE +/- 43.80, N = 3 11502 3007 6339 5408 8735 5009 20913
NVIDIA TensorRT Inference Neural Network: VGG19 - Precision: FP16 - Batch Size: 32 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: VGG19 - Precision: FP16 - Batch Size: 32 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 40 80 120 160 200 SE +/- 0.04, N = 3 SE +/- 0.05, N = 3 SE +/- 0.07, N = 3 203.96 29.83 23.94
OpenCV Benchmark OpenBenchmarking.org Seconds, Fewer Is Better OpenCV Benchmark 3.3.0 Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q ODROID-XU4 Raspberry Pi 3 Model B+ 110 220 330 440 550 SE +/- 1.57, N = 3 SE +/- 0.27, N = 3 SE +/- 5.74, N = 3 SE +/- 5.31, N = 3 128.00 296.00 493.00 520.70 2.74 1. (CXX) g++ options: -std=c++11 -rdynamic
NVIDIA TensorRT Inference Neural Network: VGG16 - Precision: INT8 - Batch Size: 32 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: VGG16 - Precision: INT8 - Batch Size: 32 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 100 200 300 400 500 SE +/- 0.10, N = 3 SE +/- 0.05, N = 3 SE +/- 0.01, N = 3 475.08 19.91 15.79
NVIDIA TensorRT Inference Neural Network: ResNet152 - Precision: FP16 - Batch Size: 32 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: ResNet152 - Precision: FP16 - Batch Size: 32 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 60 120 180 240 300 SE +/- 0.26, N = 3 SE +/- 0.07, N = 3 SE +/- 0.10, N = 3 259.82 41.91 32.67
NVIDIA TensorRT Inference Neural Network: VGG16 - Precision: FP16 - Batch Size: 32 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: VGG16 - Precision: FP16 - Batch Size: 32 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 50 100 150 200 250 SE +/- 0.12, N = 3 SE +/- 0.31, N = 3 SE +/- 0.18, N = 3 247.95 36.87 29.83
Zstd Compression Compressing ubuntu-16.04.3-server-i386.img, Compression Level 19 OpenBenchmarking.org Seconds, Fewer Is Better Zstd Compression 1.3.4 Compressing ubuntu-16.04.3-server-i386.img, Compression Level 19 ASUS TinkerBoard Jetson AGX Xavier Jetson TX1 Max-P Jetson TX2 Max-P Jetson TX2 Max-Q Raspberry Pi 3 Model B+ 110 220 330 440 550 SE +/- 2.16, N = 3 SE +/- 0.91, N = 3 SE +/- 0.42, N = 3 SE +/- 0.29, N = 3 SE +/- 1.02, N = 3 SE +/- 1.03, N = 3 496.62 80.06 145.80 144.97 253.80 342.23 1. (CC) gcc options: -O3 -pthread -lz -llzma
FLAC Audio Encoding WAV To FLAC OpenBenchmarking.org Seconds, Fewer Is Better FLAC Audio Encoding 1.3.2 WAV To FLAC ASUS TinkerBoard Jetson AGX Xavier Jetson TX1 Max-P Jetson TX2 Max-P Jetson TX2 Max-Q ODROID-XU4 Raspberry Pi 3 Model B+ 70 140 210 280 350 SE +/- 2.51, N = 5 SE +/- 0.61, N = 5 SE +/- 0.74, N = 5 SE +/- 0.15, N = 5 SE +/- 0.18, N = 5 SE +/- 0.31, N = 5 SE +/- 0.98, N = 5 279.05 54.47 79.20 65.07 104.28 97.03 339.53 1. (CXX) g++ options: -O2 -fvisibility=hidden -logg -lm
NVIDIA TensorRT Inference Neural Network: ResNet50 - Precision: FP16 - Batch Size: 4 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: ResNet50 - Precision: FP16 - Batch Size: 4 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 120 240 360 480 600 SE +/- 0.03, N = 3 SE +/- 1.32, N = 12 SE +/- 1.10, N = 12 547.50 92.28 72.01
NVIDIA TensorRT Inference Neural Network: ResNet152 - Precision: INT8 - Batch Size: 4 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: ResNet152 - Precision: INT8 - Batch Size: 4 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 80 160 240 320 400 SE +/- 1.59, N = 3 SE +/- 0.14, N = 3 SE +/- 0.15, N = 3 372.73 18.29 14.50
NVIDIA TensorRT Inference Neural Network: VGG16 - Precision: INT8 - Batch Size: 4 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: VGG16 - Precision: INT8 - Batch Size: 4 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 70 140 210 280 350 SE +/- 0.46, N = 3 SE +/- 0.25, N = 6 SE +/- 0.20, N = 5 303.78 17.56 14.24
NVIDIA TensorRT Inference Neural Network: ResNet152 - Precision: FP16 - Batch Size: 4 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: ResNet152 - Precision: FP16 - Batch Size: 4 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 50 100 150 200 250 SE +/- 0.22, N = 3 SE +/- 0.36, N = 3 SE +/- 0.34, N = 3 224.19 35.11 27.34
7-Zip Compression Compress Speed Test OpenBenchmarking.org MIPS, More Is Better 7-Zip Compression 16.02 Compress Speed Test ASUS TinkerBoard Jetson AGX Xavier Jetson TX1 Max-P Jetson TX2 Max-P Jetson TX2 Max-Q ODROID-XU4 Raspberry Pi 3 Model B+ 4K 8K 12K 16K 20K SE +/- 34.93, N = 3 SE +/- 274.18, N = 12 SE +/- 13.43, N = 3 SE +/- 20.85, N = 3 SE +/- 13.05, N = 3 SE +/- 89.16, N = 12 SE +/- 23.74, N = 11 2836 19212 4508 5593 3294 4120 2013 1. (CXX) g++ options: -pipe -lpthread
NVIDIA TensorRT Inference Neural Network: VGG19 - Precision: INT8 - Batch Size: 4 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: VGG19 - Precision: INT8 - Batch Size: 4 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 60 120 180 240 300 SE +/- 0.20, N = 3 SE +/- 0.25, N = 4 SE +/- 0.23, N = 3 265.81 14.32 11.45
NVIDIA TensorRT Inference Neural Network: ResNet50 - Precision: INT8 - Batch Size: 32 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: ResNet50 - Precision: INT8 - Batch Size: 32 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 300 600 900 1200 1500 SE +/- 0.25, N = 3 SE +/- 0.04, N = 3 SE +/- 0.08, N = 3 1215.08 59.69 47.15
GLmark2 Resolution: 1920 x 1080 OpenBenchmarking.org Score, More Is Better GLmark2 Resolution: 1920 x 1080 Jetson AGX Xavier ODROID-XU4 600 1200 1800 2400 3000 2876 1
Tesseract OCR Time To OCR 7 Images OpenBenchmarking.org Seconds, Fewer Is Better Tesseract OCR 4.0.0-beta.1 Time To OCR 7 Images Jetson AGX Xavier ODROID-XU4 40 80 120 160 200 SE +/- 0.89, N = 3 SE +/- 1.38, N = 3 71.94 180.66
NVIDIA TensorRT Inference Neural Network: GoogleNet - Precision: FP16 - Batch Size: 32 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: GoogleNet - Precision: FP16 - Batch Size: 32 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 200 400 600 800 1000 SE +/- 0.21, N = 3 SE +/- 4.50, N = 3 SE +/- 2.17, N = 8 1006 233 179
NVIDIA TensorRT Inference Neural Network: VGG16 - Precision: FP16 - Batch Size: 4 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: VGG16 - Precision: FP16 - Batch Size: 4 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 50 100 150 200 250 SE +/- 0.10, N = 3 SE +/- 0.50, N = 4 SE +/- 0.13, N = 3 208.76 32.64 25.99
NVIDIA TensorRT Inference Neural Network: VGG19 - Precision: FP16 - Batch Size: 4 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: VGG19 - Precision: FP16 - Batch Size: 4 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 40 80 120 160 200 SE +/- 0.50, N = 3 SE +/- 0.38, N = 3 SE +/- 0.34, N = 3 172.50 26.56 21.04
NVIDIA TensorRT Inference Neural Network: GoogleNet - Precision: INT8 - Batch Size: 32 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: GoogleNet - Precision: INT8 - Batch Size: 32 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 400 800 1200 1600 2000 SE +/- 8.72, N = 3 SE +/- 0.74, N = 3 SE +/- 0.07, N = 3 1693 130 104
NVIDIA TensorRT Inference Neural Network: GoogleNet - Precision: FP16 - Batch Size: 4 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: GoogleNet - Precision: FP16 - Batch Size: 4 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 200 400 600 800 1000 SE +/- 2.48, N = 3 SE +/- 2.27, N = 3 SE +/- 1.90, N = 12 796 197 156
NVIDIA TensorRT Inference Neural Network: ResNet50 - Precision: FP16 - Batch Size: 32 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: ResNet50 - Precision: FP16 - Batch Size: 32 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 140 280 420 560 700 SE +/- 1.23, N = 3 SE +/- 1.22, N = 3 SE +/- 0.86, N = 3 636.00 111.00 86.08
NVIDIA TensorRT Inference Neural Network: AlexNet - Precision: FP16 - Batch Size: 32 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: AlexNet - Precision: FP16 - Batch Size: 32 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 400 800 1200 1600 2000 SE +/- 2.07, N = 3 SE +/- 7.68, N = 12 SE +/- 2.82, N = 3 2038 462 374
NVIDIA TensorRT Inference Neural Network: AlexNet - Precision: FP16 - Batch Size: 4 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: AlexNet - Precision: FP16 - Batch Size: 4 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 300 600 900 1200 1500 SE +/- 1.82, N = 3 SE +/- 7.77, N = 12 SE +/- 3.03, N = 6 1200 264 216
NVIDIA TensorRT Inference Neural Network: ResNet50 - Precision: INT8 - Batch Size: 4 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: ResNet50 - Precision: INT8 - Batch Size: 4 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 200 400 600 800 1000 SE +/- 1.86, N = 3 SE +/- 0.79, N = 4 SE +/- 0.64, N = 3 902.78 49.97 39.15
NVIDIA TensorRT Inference Neural Network: GoogleNet - Precision: INT8 - Batch Size: 4 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: GoogleNet - Precision: INT8 - Batch Size: 4 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 200 400 600 800 1000 SE +/- 4.31, N = 3 SE +/- 1.65, N = 3 SE +/- 1.32, N = 3 1146.00 113.00 88.88
NVIDIA TensorRT Inference Neural Network: AlexNet - Precision: INT8 - Batch Size: 4 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: AlexNet - Precision: INT8 - Batch Size: 4 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 200 400 600 800 1000 SE +/- 2.59, N = 3 SE +/- 2.79, N = 5 SE +/- 0.91, N = 3 1143 184 148
NVIDIA TensorRT Inference Neural Network: AlexNet - Precision: INT8 - Batch Size: 32 - DLA Cores: Disabled OpenBenchmarking.org Images Per Second, More Is Better NVIDIA TensorRT Inference Neural Network: AlexNet - Precision: INT8 - Batch Size: 32 - DLA Cores: Disabled Jetson AGX Xavier Jetson TX2 Max-P Jetson TX2 Max-Q 700 1400 2100 2800 3500 SE +/- 1.06, N = 3 SE +/- 0.52, N = 3 SE +/- 1.39, N = 3 3143 301 237
LeelaChessZero Backend: CUDA + cuDNN OpenBenchmarking.org Nodes Per Second, More Is Better LeelaChessZero 0.20.1 Backend: CUDA + cuDNN Jetson AGX Xavier 200 400 600 800 1000 SE +/- 6.14, N = 3 952.89 1. (CXX) g++ options: -lpthread -lz
LeelaChessZero Backend: CUDA + cuDNN FP16 OpenBenchmarking.org Nodes Per Second, More Is Better LeelaChessZero 0.20.1 Backend: CUDA + cuDNN FP16 Jetson AGX Xavier 500 1000 1500 2000 2500 SE +/- 7.60, N = 3 2515.01 1. (CXX) g++ options: -lpthread -lz
Phoronix Test Suite v10.8.4