-
Notifications
You must be signed in to change notification settings - Fork 2.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Failed to utilise CUDA with TRT Engine when running on Jetson AGX Orin (ONNX->TRT, Transformer) #2997
Comments
It's because your model contains operators that can not be capture by cuda graph, like loops or if-condition operators. see https://docs.nvidia.com/deeplearning/tensorrt/developer-guide/index.html#cuda-graphs To get the better performance, I have a few suggestions: 1. use the latest TRT, which has better optimization and new feature. 2. use |
@niqbal996 Hello, How could you transfer DINO .pth to onnx format? Thanks. |
Hey @j0987834204, |
@niqbal996 |
@niqbal996 @zerollzeng |
Description
I am trying to convert a DINO object detector Transformer trained with a custom dataset model to a TensorRT engine with any precision. I am using trtexec for that. The engine file is generated but it gives the following error:
I can perform inference with a python script but I think it is not using the GPU for that and can only run at 3 FPS on Jetson Orin. I am expecting roughly at least 10 FPS on the Jetson Orin.
Any ideas what might be causing this issue and how I can solve that?
I used the following command for conversion:
Verbose output logs
Environment
TensorRT Version: 8.5.2
NVIDIA GPU: Jetson AGX ORIN
NVIDIA Driver Version: L4T 35.3.1
CUDA Version: 11.4.315
CUDNN Version: 8.6.0.166
Operating System: Ubuntu 20.04 LTS
Python Version (if applicable): 3.8
PyTorch Version (if applicable):
Container (if so, version): nvcr.io/nvidia/l4t-tensorrt:r8.5.2.2-devel
Relevant Files
Model link:
The model onnx file and the full verbose log output file can be downloaded at the following link: drive
Steps To Reproduce
Commands or scripts:
trtexec --onnx=dino_simp.onnx --int8 --useCudaGraph --verbose --saveEngine=dino_last.trt --workspace=20000
Have you tried the latest release?: The latest TensorRT release I can only try on my laptop but the corresponding Jetpack release is not yet available to be installed on the Jetson Orin. #2949
Can this model run on other frameworks? I can do inference with ONNX runtime on my model. I also tried to convert the same model on my laptop and it works without any issues. I can do inference with about 14 FPS which is expected.
Thank you for looking into this.
The text was updated successfully, but these errors were encountered: