Jetson AGX Xavier vs. Jetson TX2 TensorRT

NVIDIA Jetson TensorRT inference benchmarks by Michael Larabel for a future article on Phoronix.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 1812240-SP-XAVIER80657
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Additional Graphs

Show Perf Per Core/Thread Calculation Graphs Where Applicable
Show Perf Per Clock Calculation Graphs Where Applicable

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
Jetson AGX Xavier
December 23 2018
  6 Hours, 30 Minutes
Jetson TX2
December 24 2018
  6 Hours, 26 Minutes
Invert Hiding All Results Option
  6 Hours, 28 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


Jetson AGX Xavier vs. Jetson TX2 TensorRT - Phoronix Test Suite

Jetson AGX Xavier vs. Jetson TX2 TensorRT

NVIDIA Jetson TensorRT inference benchmarks by Michael Larabel for a future article on Phoronix.

HTML result view exported from: https://openbenchmarking.org/result/1812240-SP-XAVIER80657&obr_sor=y&obr_rro=y&rdt&grs.

Jetson AGX Xavier vs. Jetson TX2 TensorRTProcessorMotherboardMemoryDiskGraphicsMonitorOSKernelDesktopDisplay ServerDisplay DriverOpenGLVulkanCompilerFile-SystemScreen ResolutionJetson AGX XavierJetson TX2ARMv8 rev 0 @ 2.27GHz (8 Cores)jetson-xavier16384MB31GB HBG4a2NVIDIA Tegra XavierASUS VP28UUbuntu 18.044.9.108-tegra (aarch64)Unity 7.5.0X Server 1.19.6NVIDIA 31.0.24.6.01.1.76GCC 7.3.0 + CUDA 10.0ext41920x1080ARMv8 rev 3 @ 2.04GHz (4 Cores / 6 Threads)quill8192MB31GB 032G34NVIDIA Tegra X2VE228Ubuntu 16.044.4.38-tegra (aarch64)Unity 7.4.0X Server 1.18.4NVIDIA 28.2.14.5.0GCC 5.4.0 20160609 + CUDA 9.0OpenBenchmarking.orgProcessor Details- Scaling Governor: tegra_cpufreq schedutil

Jetson AGX Xavier vs. Jetson TX2 TensorRTtensorrt-inference: VGG19 - FP16 - 8tensorrt-inference: VGG19 - FP16 - 32tensorrt-inference: VGG16 - FP16 - 32tensorrt-inference: VGG19 - FP16 - 4tensorrt-inference: VGG16 - FP16 - 8tensorrt-inference: ResNet152 - FP16 - 8tensorrt-inference: VGG16 - FP16 - 16tensorrt-inference: ResNet152 - FP16 - 4tensorrt-inference: VGG16 - FP16 - 4tensorrt-inference: ResNet152 - FP16 - 32tensorrt-inference: ResNet50 - FP16 - 8tensorrt-inference: ResNet50 - FP16 - 4tensorrt-inference: ResNet50 - FP16 - 16tensorrt-inference: ResNet50 - FP16 - 32tensorrt-inference: GoogleNet - FP16 - 8tensorrt-inference: GoogleNet - FP16 - 32tensorrt-inference: AlexNet - FP16 - 32tensorrt-inference: VGG19 - INT8 - 32tensorrt-inference: VGG16 - INT8 - 32tensorrt-inference: VGG19 - INT8 - 16tensorrt-inference: ResNet152 - INT8 - 32tensorrt-inference: ResNet152 - INT8 - 16tensorrt-inference: ResNet152 - INT8 - 8tensorrt-inference: ResNet50 - INT8 - 32tensorrt-inference: ResNet152 - INT8 - 4tensorrt-inference: ResNet50 - INT8 - 16tensorrt-inference: VGG19 - INT8 - 8tensorrt-inference: VGG19 - INT8 - 4tensorrt-inference: ResNet50 - INT8 - 4tensorrt-inference: VGG16 - INT8 - 8tensorrt-inference: VGG16 - INT8 - 4tensorrt-inference: GoogleNet - INT8 - 32glmark2: 1920 x 1080tensorrt-inference: ResNet152 - FP16 - 16tensorrt-inference: GoogleNet - INT8 - 16tensorrt-inference: GoogleNet - FP16 - 16tensorrt-inference: GoogleNet - INT8 - 8tensorrt-inference: GoogleNet - INT8 - 4tensorrt-inference: GoogleNet - FP16 - 4tensorrt-inference: ResNet50 - INT8 - 8tensorrt-inference: AlexNet - INT8 - 32tensorrt-inference: AlexNet - INT8 - 16tensorrt-inference: AlexNet - FP16 - 16tensorrt-inference: AlexNet - INT8 - 8tensorrt-inference: AlexNet - INT8 - 4tensorrt-inference: AlexNet - FP16 - 8tensorrt-inference: AlexNet - FP16 - 4tensorrt-inference: VGG19 - FP16 - 16tensorrt-inference: VGG16 - INT8 - 16Jetson AGX XavierJetson TX2184.43201.53246.76172.15215.68234.84228.75219.08195.45253.34582.36542.805936138639561900390.57449.96362.08485.22445.22407.011184.50350.281106.13296.94262.17865.46341.20286.6416222861224.6013408581049652546944.4626661879143512379751247799180.03381.3327.0229.5737.2426.4933.6836.7136.4435.6032.3041.8799.0593.6110611019823047215.9919.8716.3822.0520.7719.4859.4517.9757.1816.0214.6250.3919.8917.9813040.1912521811711420251.0730725837022217930026128.9720.50OpenBenchmarking.org

NVIDIA TensorRT Inference

Neural Network: VGG19 - Precision: FP16 - Batch Size: 8

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG19 - Precision: FP16 - Batch Size: 8Jetson TX2Jetson AGX Xavier4080120160200SE +/- 0.14, N = 3SE +/- 2.36, N = 327.02184.43

NVIDIA TensorRT Inference

Neural Network: VGG19 - Precision: FP16 - Batch Size: 32

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG19 - Precision: FP16 - Batch Size: 32Jetson TX2Jetson AGX Xavier4080120160200SE +/- 0.09, N = 3SE +/- 1.68, N = 329.57201.53

NVIDIA TensorRT Inference

Neural Network: VGG16 - Precision: FP16 - Batch Size: 32

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG16 - Precision: FP16 - Batch Size: 32Jetson TX2Jetson AGX Xavier50100150200250SE +/- 0.14, N = 3SE +/- 0.17, N = 337.24246.76

NVIDIA TensorRT Inference

Neural Network: VGG19 - Precision: FP16 - Batch Size: 4

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG19 - Precision: FP16 - Batch Size: 4Jetson TX2Jetson AGX Xavier4080120160200SE +/- 0.15, N = 3SE +/- 1.25, N = 326.49172.15

NVIDIA TensorRT Inference

Neural Network: VGG16 - Precision: FP16 - Batch Size: 8

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG16 - Precision: FP16 - Batch Size: 8Jetson TX2Jetson AGX Xavier50100150200250SE +/- 0.24, N = 3SE +/- 3.36, N = 533.68215.68

NVIDIA TensorRT Inference

Neural Network: ResNet152 - Precision: FP16 - Batch Size: 8

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet152 - Precision: FP16 - Batch Size: 8Jetson TX2Jetson AGX Xavier50100150200250SE +/- 0.67, N = 9SE +/- 0.36, N = 336.71234.84

NVIDIA TensorRT Inference

Neural Network: VGG16 - Precision: FP16 - Batch Size: 16

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG16 - Precision: FP16 - Batch Size: 16Jetson TX2Jetson AGX Xavier50100150200250SE +/- 0.11, N = 3SE +/- 1.63, N = 336.44228.75

NVIDIA TensorRT Inference

Neural Network: ResNet152 - Precision: FP16 - Batch Size: 4

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet152 - Precision: FP16 - Batch Size: 4Jetson TX2Jetson AGX Xavier50100150200250SE +/- 0.44, N = 3SE +/- 3.18, N = 335.60219.08

NVIDIA TensorRT Inference

Neural Network: VGG16 - Precision: FP16 - Batch Size: 4

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG16 - Precision: FP16 - Batch Size: 4Jetson TX2Jetson AGX Xavier4080120160200SE +/- 0.30, N = 3SE +/- 3.17, N = 1232.30195.45

NVIDIA TensorRT Inference

Neural Network: ResNet152 - Precision: FP16 - Batch Size: 32

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet152 - Precision: FP16 - Batch Size: 32Jetson TX2Jetson AGX Xavier60120180240300SE +/- 0.14, N = 3SE +/- 2.84, N = 341.87253.34

NVIDIA TensorRT Inference

Neural Network: ResNet50 - Precision: FP16 - Batch Size: 8

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet50 - Precision: FP16 - Batch Size: 8Jetson TX2Jetson AGX Xavier130260390520650SE +/- 1.23, N = 3SE +/- 0.24, N = 399.05582.36

NVIDIA TensorRT Inference

Neural Network: ResNet50 - Precision: FP16 - Batch Size: 4

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet50 - Precision: FP16 - Batch Size: 4Jetson TX2Jetson AGX Xavier120240360480600SE +/- 1.46, N = 3SE +/- 0.39, N = 393.61542.80

NVIDIA TensorRT Inference

Neural Network: ResNet50 - Precision: FP16 - Batch Size: 16

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet50 - Precision: FP16 - Batch Size: 16Jetson TX2Jetson AGX Xavier130260390520650SE +/- 0.59, N = 3SE +/- 7.03, N = 3106593

NVIDIA TensorRT Inference

Neural Network: ResNet50 - Precision: FP16 - Batch Size: 32

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet50 - Precision: FP16 - Batch Size: 32Jetson TX2Jetson AGX Xavier130260390520650SE +/- 1.29, N = 3SE +/- 9.12, N = 3110613

NVIDIA TensorRT Inference

Neural Network: GoogleNet - Precision: FP16 - Batch Size: 8

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: GoogleNet - Precision: FP16 - Batch Size: 8Jetson TX2Jetson AGX Xavier2004006008001000SE +/- 3.70, N = 3SE +/- 14.25, N = 12198863

NVIDIA TensorRT Inference

Neural Network: GoogleNet - Precision: FP16 - Batch Size: 32

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: GoogleNet - Precision: FP16 - Batch Size: 32Jetson TX2Jetson AGX Xavier2004006008001000SE +/- 3.59, N = 3SE +/- 14.46, N = 12230956

NVIDIA TensorRT Inference

Neural Network: AlexNet - Precision: FP16 - Batch Size: 32

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: AlexNet - Precision: FP16 - Batch Size: 32Jetson TX2Jetson AGX Xavier400800120016002000SE +/- 6.74, N = 3SE +/- 23.33, N = 34721900

NVIDIA TensorRT Inference

Neural Network: VGG19 - Precision: INT8 - Batch Size: 32

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG19 - Precision: INT8 - Batch Size: 32Jetson TX2Jetson AGX Xavier80160240320400SE +/- 0.03, N = 3SE +/- 1.67, N = 315.99390.57

NVIDIA TensorRT Inference

Neural Network: VGG16 - Precision: INT8 - Batch Size: 32

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG16 - Precision: INT8 - Batch Size: 32Jetson TX2Jetson AGX Xavier100200300400500SE +/- 0.03, N = 3SE +/- 4.97, N = 1019.87449.96

NVIDIA TensorRT Inference

Neural Network: VGG19 - Precision: INT8 - Batch Size: 16

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG19 - Precision: INT8 - Batch Size: 16Jetson TX2Jetson AGX Xavier80160240320400SE +/- 0.02, N = 3SE +/- 0.66, N = 316.38362.08

NVIDIA TensorRT Inference

Neural Network: ResNet152 - Precision: INT8 - Batch Size: 32

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet152 - Precision: INT8 - Batch Size: 32Jetson TX2Jetson AGX Xavier110220330440550SE +/- 0.03, N = 3SE +/- 1.47, N = 322.05485.22

NVIDIA TensorRT Inference

Neural Network: ResNet152 - Precision: INT8 - Batch Size: 16

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet152 - Precision: INT8 - Batch Size: 16Jetson TX2Jetson AGX Xavier100200300400500SE +/- 0.09, N = 3SE +/- 4.04, N = 320.77445.22

NVIDIA TensorRT Inference

Neural Network: ResNet152 - Precision: INT8 - Batch Size: 8

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet152 - Precision: INT8 - Batch Size: 8Jetson TX2Jetson AGX Xavier90180270360450SE +/- 0.27, N = 3SE +/- 6.98, N = 319.48407.01

NVIDIA TensorRT Inference

Neural Network: ResNet50 - Precision: INT8 - Batch Size: 32

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet50 - Precision: INT8 - Batch Size: 32Jetson TX2Jetson AGX Xavier30060090012001500SE +/- 0.19, N = 3SE +/- 6.54, N = 359.451184.50

NVIDIA TensorRT Inference

Neural Network: ResNet152 - Precision: INT8 - Batch Size: 4

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet152 - Precision: INT8 - Batch Size: 4Jetson TX2Jetson AGX Xavier80160240320400SE +/- 0.19, N = 3SE +/- 5.48, N = 317.97350.28

NVIDIA TensorRT Inference

Neural Network: ResNet50 - Precision: INT8 - Batch Size: 16

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet50 - Precision: INT8 - Batch Size: 16Jetson TX2Jetson AGX Xavier2004006008001000SE +/- 0.10, N = 3SE +/- 11.53, N = 1257.181106.13

NVIDIA TensorRT Inference

Neural Network: VGG19 - Precision: INT8 - Batch Size: 8

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG19 - Precision: INT8 - Batch Size: 8Jetson TX2Jetson AGX Xavier60120180240300SE +/- 0.06, N = 3SE +/- 1.42, N = 316.02296.94

NVIDIA TensorRT Inference

Neural Network: VGG19 - Precision: INT8 - Batch Size: 4

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG19 - Precision: INT8 - Batch Size: 4Jetson TX2Jetson AGX Xavier60120180240300SE +/- 0.10, N = 3SE +/- 0.96, N = 314.62262.17

NVIDIA TensorRT Inference

Neural Network: ResNet50 - Precision: INT8 - Batch Size: 4

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet50 - Precision: INT8 - Batch Size: 4Jetson TX2Jetson AGX Xavier2004006008001000SE +/- 0.64, N = 3SE +/- 14.20, N = 350.39865.46

NVIDIA TensorRT Inference

Neural Network: VGG16 - Precision: INT8 - Batch Size: 8

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG16 - Precision: INT8 - Batch Size: 8Jetson TX2Jetson AGX Xavier70140210280350SE +/- 0.05, N = 3SE +/- 1.08, N = 319.89341.20

NVIDIA TensorRT Inference

Neural Network: VGG16 - Precision: INT8 - Batch Size: 4

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG16 - Precision: INT8 - Batch Size: 4Jetson TX2Jetson AGX Xavier60120180240300SE +/- 0.06, N = 3SE +/- 3.98, N = 317.98286.64

NVIDIA TensorRT Inference

Neural Network: GoogleNet - Precision: INT8 - Batch Size: 32

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: GoogleNet - Precision: INT8 - Batch Size: 32Jetson TX2Jetson AGX Xavier30060090012001500SE +/- 0.91, N = 3SE +/- 5.04, N = 31301622

GLmark2

Resolution: 1920 x 1080

OpenBenchmarking.orgScore, More Is BetterGLmark2Resolution: 1920 x 1080Jetson AGX Xavier60012001800240030002861

NVIDIA TensorRT Inference

Neural Network: ResNet152 - Precision: FP16 - Batch Size: 16

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet152 - Precision: FP16 - Batch Size: 16Jetson TX2Jetson AGX Xavier50100150200250SE +/- 0.17, N = 3SE +/- 15.50, N = 940.19224.60

NVIDIA TensorRT Inference

Neural Network: GoogleNet - Precision: INT8 - Batch Size: 16

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: GoogleNet - Precision: INT8 - Batch Size: 16Jetson TX2Jetson AGX Xavier30060090012001500SE +/- 1.16, N = 3SE +/- 152.29, N = 91251340

NVIDIA TensorRT Inference

Neural Network: GoogleNet - Precision: FP16 - Batch Size: 16

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: GoogleNet - Precision: FP16 - Batch Size: 16Jetson TX2Jetson AGX Xavier2004006008001000SE +/- 3.60, N = 3SE +/- 55.00, N = 9218858

NVIDIA TensorRT Inference

Neural Network: GoogleNet - Precision: INT8 - Batch Size: 8

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: GoogleNet - Precision: INT8 - Batch Size: 8Jetson TX2Jetson AGX Xavier2004006008001000SE +/- 2.12, N = 3SE +/- 121.56, N = 101171049

NVIDIA TensorRT Inference

Neural Network: GoogleNet - Precision: INT8 - Batch Size: 4

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: GoogleNet - Precision: INT8 - Batch Size: 4Jetson TX2Jetson AGX Xavier140280420560700SE +/- 2.00, N = 3SE +/- 140.60, N = 12114652

NVIDIA TensorRT Inference

Neural Network: GoogleNet - Precision: FP16 - Batch Size: 4

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: GoogleNet - Precision: FP16 - Batch Size: 4Jetson TX2Jetson AGX Xavier120240360480600SE +/- 0.88, N = 3SE +/- 96.56, N = 9202546

NVIDIA TensorRT Inference

Neural Network: ResNet50 - Precision: INT8 - Batch Size: 8

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: ResNet50 - Precision: INT8 - Batch Size: 8Jetson TX2Jetson AGX Xavier2004006008001000SE +/- 0.54, N = 3SE +/- 40.28, N = 1251.07944.46

NVIDIA TensorRT Inference

Neural Network: AlexNet - Precision: INT8 - Batch Size: 32

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: AlexNet - Precision: INT8 - Batch Size: 32Jetson TX2Jetson AGX Xavier6001200180024003000SE +/- 0.88, N = 3SE +/- 248.85, N = 93072666

NVIDIA TensorRT Inference

Neural Network: AlexNet - Precision: INT8 - Batch Size: 16

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: AlexNet - Precision: INT8 - Batch Size: 16Jetson TX2Jetson AGX Xavier400800120016002000SE +/- 3.45, N = 3SE +/- 91.41, N = 122581879

NVIDIA TensorRT Inference

Neural Network: AlexNet - Precision: FP16 - Batch Size: 16

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: AlexNet - Precision: FP16 - Batch Size: 16Jetson TX2Jetson AGX Xavier30060090012001500SE +/- 6.40, N = 12SE +/- 89.56, N = 93701435

NVIDIA TensorRT Inference

Neural Network: AlexNet - Precision: INT8 - Batch Size: 8

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: AlexNet - Precision: INT8 - Batch Size: 8Jetson TX2Jetson AGX Xavier30060090012001500SE +/- 3.23, N = 3SE +/- 99.61, N = 122221237

NVIDIA TensorRT Inference

Neural Network: AlexNet - Precision: INT8 - Batch Size: 4

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: AlexNet - Precision: INT8 - Batch Size: 4Jetson TX2Jetson AGX Xavier2004006008001000SE +/- 2.69, N = 4SE +/- 55.83, N = 12179975

NVIDIA TensorRT Inference

Neural Network: AlexNet - Precision: FP16 - Batch Size: 8

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: AlexNet - Precision: FP16 - Batch Size: 8Jetson TX2Jetson AGX Xavier30060090012001500SE +/- 7.60, N = 12SE +/- 45.66, N = 123001247

NVIDIA TensorRT Inference

Neural Network: AlexNet - Precision: FP16 - Batch Size: 4

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: AlexNet - Precision: FP16 - Batch Size: 4Jetson TX2Jetson AGX Xavier2004006008001000SE +/- 5.89, N = 12SE +/- 97.79, N = 9261799

NVIDIA TensorRT Inference

Neural Network: VGG19 - Precision: FP16 - Batch Size: 16

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG19 - Precision: FP16 - Batch Size: 16Jetson TX2Jetson AGX Xavier4080120160200SE +/- 0.11, N = 3SE +/- 11.67, N = 1028.97180.03

NVIDIA TensorRT Inference

Neural Network: VGG16 - Precision: INT8 - Batch Size: 16

OpenBenchmarking.orgImages Per Second, More Is BetterNVIDIA TensorRT InferenceNeural Network: VGG16 - Precision: INT8 - Batch Size: 16Jetson TX2Jetson AGX Xavier80160240320400SE +/- 0.03, N = 3SE +/- 10.09, N = 1220.50381.33


Phoronix Test Suite v10.8.4