Skip to content

Supported hardware

TargetHardwareMemoryPrecision
orin-nanoJetson Orin Nano8 GBfp16
orinJetson Orin32 GBfp16
orin-64Jetson Orin 6464 GBfp16
thorJetson Thor128 GBfp8
desktopRTX / A10040 GBfp16

reflex inspect targets lists current profiles.

ArchitectureComputeStatusNotes
Ampere (RTX 30-series, A10G, A100)sm_8.0–8.6SupportedTested on Modal A10G + A100, RTX 4090
Ada Lovelace (RTX 40-series, L4)sm_8.9Supported
Hopper (H100, H200)sm_9.0Supported
Jetson Orin (Orin Nano / NX / AGX)sm_8.7SupportedJetPack 5.x or 6.x
Jetson Thorsm_10.xUntestedSame Blackwell silicon as desktop, but ORT-bundled CUDA EP needs Blackwell support
Blackwell desktop (RTX 5090, RTX PRO 6000, B200, GB200)sm_10.0Not yet supportedORT’s bundled cuBLAS/cuDNN don’t ship sm_100 kernels
Older NVIDIA (Turing RTX 20, GTX 16)sm_7.5Best-effortShould work but not in CI matrix
Pre-Tensor-Core (Maxwell Jetson Nano 4 GB, GTX 9-series)sm_5.xNot supportedNVIDIA EOL’d this hardware at JetPack 4.6 (Python 3.6)

reflex-vla[serve,gpu] (v0.7+) uses ONNX Runtime’s TensorRT execution provider out of the box. Measured on Modal A10G (Ampere, sm_8.6) on 2026-04-29 against SmolVLA monolithic (5 warmup + 20 measured forward passes, batch=1):

ProviderMean latencyp95
CUDAExecutionProvider (ORT-CUDA fallback)108.11 ms108.68 ms
TensorrtExecutionProvider (default in v0.7+)19.49 ms19.71 ms

5.55× faster. The win comes from TensorRT’s FP16 kernels + engine fusion. Older releases silently fell back to ORT-CUDA on most installs because libnvinfer.so.10 and CUDA libs weren’t on LD_LIBRARY_PATH — v0.7’s [serve,gpu] extras pull tensorrt>=10 and reflex patches LD_LIBRARY_PATH automatically at import.

Terminal window
reflex doctor

Look for green ✓ on:

  • TensorRT runtime (libnvinfer.so.10) — loadable
  • CUDA cuBLAS (libcublas.so.12) — loadable
  • CUDA cuDNN (libcudnn.so.9) — loadable
  • ORT-TRT EP active — session created with TRT EP in active providers

If any are red ✗, the remediation hint says exactly which pip install to run. The most common cause is using [serve,gpu-min] or an older release that didn’t pull tensorrt automatically.