-
Notifications
You must be signed in to change notification settings - Fork 351
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
🐛 [Bug] Cannot export models to TensorRT with int8 quantization #1222
Comments
This issue has not seen activity for 90 days, Remove stale label or comment or this will be closed in 10 days |
I encountered the same issue , |
I use the example in tutorial https://pytorch.org/TensorRT/tutorials/use_from_pytorch.html#, same problem occurred. |
+1 have the same problem! |
+1 |
@peri044 Could you take a look at the matter, please? |
I have the same issue, any update in this? |
@AhmetHamzaEmra Can you provide a repro of the error and the error message ? 22.06 container is quite old and you should probably try with the main branch |
|
I have the same issue, any update in this? |
This issue has not seen activity for 90 days, Remove stale label or comment or this will be closed in 10 days |
Any update on the uint8 export? |
Is int8 quantization now working? |
I saw same error message as this page. |
Any update on the quantization the model to int8 format ? |
🐛 Describe the bug
I'm trying to convert a resnet18 to TensorRT. It works fine when setting
enabled_precisions
totorch.float
and totorch.float16
. It doesn't work withtorch.int8
.When using the model exported with
torch.jit.script
, the error is the following:When using the model exported with
torch.jit.trace
, the program exit with:I'm using the nvidia container
nvcr.io/nvidia/pytorch:22.06-py3
.Edit:
If I convert the
nn.Module
I get the same error as using the scripted model:Versions
PyTorch version: 1.13.0a0+340c412
Is debug build: False
CUDA used to build PyTorch: 11.7
ROCM used to build PyTorch: N/A
OS: Ubuntu 20.04.4 LTS (x86_64)
GCC version: (Ubuntu 9.4.0-1ubuntu1~20.04.1) 9.4.0
Clang version: Could not collect
CMake version: version 3.23.2
Libc version: glibc-2.31
Python version: 3.8.13 | packaged by conda-forge | (default, Mar 25 2022, 06:04:10) [GCC 10.3.0] (64-bit runtime)
Python platform: Linux-5.13.0-52-generic-x86_64-with-glibc2.10
Is CUDA available: True
CUDA runtime version: 11.7.99
GPU models and configuration: GPU 0: NVIDIA GeForce RTX 2080 Ti
Nvidia driver version: 510.73.05
cuDNN version: Probably one of the following:
/usr/lib/x86_64-linux-gnu/libcudnn.so.8.4.1
/usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.4.1
/usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.4.1
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.4.1
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.4.1
/usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.4.1
/usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.4.1
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
Versions of relevant libraries:
[pip3] numpy==1.22.4
[pip3] pytorch-quantization==2.1.2
[pip3] torch==1.13.0a0+340c412
[pip3] torch-tensorrt==1.1.0a0
[pip3] torchtext==0.13.0a0
[pip3] torchvision==0.13.0a0
[conda] mkl 2020.4 h726a3e6_304 conda-forge
[conda] mkl-include 2020.4 h726a3e6_304 conda-forge
[conda] numpy 1.22.4 py38h99721a1_0 conda-forge
[conda] pytorch-quantization 2.1.2 pypi_0 pypi
[conda] torch 1.13.0a0+340c412 pypi_0 pypi
[conda] torch-tensorrt 1.1.0a0 pypi_0 pypi
[conda] torchtext 0.13.0a0 pypi_0 pypi
[conda] torchvision 0.13.0a0 pypi_0 pypi
cc @jerryzh168 @jianyuh @raghuramank100 @jamesr66a @vkuzo
The text was updated successfully, but these errors were encountered: