-
Notifications
You must be signed in to change notification settings - Fork 2.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Output mismatch of torch.add due to an intermediate result output when running on GPU #3452
Labels
internal-bug-tracked
Tracked internally, will be fixed in a future release.
triaged
Issue has been triaged by maintainers
Comments
I can repro in TRT 9.1.0.post12.dev4. @zerollzeng |
zerollzeng
added
triaged
Issue has been triaged by maintainers
internal-bug-tracked
Tracked internally, will be fixed in a future release.
labels
Nov 18, 2023
I'll file internal bug to track this, we need to debug further to decide whether this is a bug or not. maybe the diff is caused by layer fusion
|
Filed internal bug 4383767 for this. |
Diff is expected for FP16, not a bug, close this. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Labels
internal-bug-tracked
Tracked internally, will be fixed in a future release.
triaged
Issue has been triaged by maintainers
Description
When outputting an intermediate result of
torch.neg
in this model:New:
The original output of
torch.add
is expected to be the same for the same input in this 2 graphs. However, it mismatched bewteew the 2 models.Environment
TensorRT Version: 8.6.1.post1
NVIDIA GPU: RTX 1660
NVIDIA Driver Version: 525.147.05
CUDA Version: 12.0
CUDNN Version: 8.9.4.25
Operating System: Ubuntu 22.04.3 LTS (x86_64)
Python Version (if applicable): 3.10.12
Tensorflow Version (if applicable): 2.13.0
PyTorch Version (if applicable): 2.1.0+cu118
Relevant Files
Model link:
models.zip
Input data file:
input_data.zip
Steps To Reproduce
Script:
Steps to repro:
Output assertion:
Have you tried the latest release?: No
Can this model run on other frameworks? For example run ONNX model with ONNXRuntime (
polygraphy run <model.onnx> --onnxrt
): YesThe text was updated successfully, but these errors were encountered: