We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
tensorrt cannot work with int8 precision when using on NanoGPT like (bark text-to-audio) model.
Steps to reproduce the behavior:
inp1, inp2 = get_sample_input() inp1 = torch.unsqueeze(inp1, 0) traced_model = torch.jit.trace(model, example_inputs=[inp1, inp2]) batch_size = 1 trt_model = torch_tensorrt.compile( traced_model, inputs = [ torch_tensorrt.Input((batch_size,1), dtype=torch.long), torch_tensorrt.Input((batch_size, 1024, 8), dtype=torch.long)], enabled_precisions = { torch.int8}, workspace_size=20000000000, truncate_long_and_double = True )
and I am getting following error:
Segmentation fault
When I run the same code on float32 or half precion everyting works fine. only happens when its int8
Build information about Torch-TensorRT can be found by turning on debug messages
conda
pip
libtorch
The text was updated successfully, but these errors were encountered:
I have the same issue, any update in this?
Sorry, something went wrong.
I am waiting for an update too 😅
This issue has not seen activity for 90 days, Remove stale label or comment or this will be closed in 10 days
Why are you guys keep closing the issue? We are still waitign for an update ?????
peri044
No branches or pull requests
Bug Description
tensorrt cannot work with int8 precision when using on NanoGPT like (bark text-to-audio) model.
To Reproduce
Steps to reproduce the behavior:
and I am getting following error:
Expected behavior
When I run the same code on float32 or half precion everyting works fine. only happens when its int8
Environment
conda
,pip
,libtorch
, source): pipAdditional context
The text was updated successfully, but these errors were encountered: