Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The trt model is used on different graphics cards #4259

Open
wahaha opened this issue Nov 23, 2024 · 3 comments
Open

The trt model is used on different graphics cards #4259

wahaha opened this issue Nov 23, 2024 · 3 comments
Labels
triaged Issue has been triaged by maintainers

Comments

@wahaha
Copy link

wahaha commented Nov 23, 2024

Hello, is there any way that a trt model running on a 30 series graphics card can be inferred on a 20 series graphics card? For example, modify a parameter? Looking forward to your reply!

@lix19937
Copy link

You has better to build and run the engine a the same machine(gpu arch).

@wahaha
Copy link
Author

wahaha commented Nov 25, 2024

Thank your very much for your answer! I want to ask is there no way to build and run engine at the different machine(gpu arch), ex. build the engine on the 30-series gpu arch and then run the engine on the 20-series gpu arch? Or is build and run engine at the different machine(gpu arch) prone to other problems? Looking forward to your reply!

@lix19937
Copy link

If GPU clock speeds differ between engine serialization and runtime systems, the tactics chosen by the serialization system may not be optimal for the runtime system and may incur some performance degradation.

If it is impossible to build a TensorRT engine for each individual type of GPU, you can select several GPUs to build engines with and run the engine on different GPUs with the same architecture. For example, among the NVIDIA RTX 40xx GPUs, you can build an engine with RTX 4090 and an engine with RTX 4070. At runtime, you can use the RTX 4090 engine on an RTX 4080 GPU and the 4070 engine on all smaller GPUs. In most cases, the engine will run without functional issues and with only a small performance drop compared to running the engine built with the same GPU.

you can ref https://docs.nvidia.com/deeplearning/tensorrt/developer-guide/index.html#hardware-compat and https://docs.nvidia.com/deeplearning/tensorrt/developer-guide/index.html#compatibility-checks

@poweiw poweiw added the triaged Issue has been triaged by maintainers label Dec 2, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
triaged Issue has been triaged by maintainers
Projects
None yet
Development

No branches or pull requests

3 participants