-
Notifications
You must be signed in to change notification settings - Fork 473
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Help wanted] Support TensorRT #40
Comments
I would like take on this.
|
Hi @csukuangfj , @yuekaizhang Observed that currently only CUDA EP support is there and TensorRT EP support is not there for onnxruntime. |
We don't have a plan to support it in the near future. Would you like to contribute? |
I tried adding triggering onnxruntime's tensorrt ep for zipfromer but the model performance was very bad, |
Hi @csukuangfj,
Thank you |
Could you create a new config for tensorrt and add this config as a member field of You can set the default values of this config as the one used in sherpa-onnx/sherpa-onnx/csrc/session.cc Lines 137 to 150 in b714817
|
yes, I will send the PR for configs separately in some time. |
Current perf Trt Vs Cuda Tensorrt Cuda Apart from this, with Trt there is a huge session creation time. |
May I know the results for CPU provider if you have? Also, could you explain why there are three lines for each block? e.g. 0.535651 ms 0.187492 ms 0.187698 ms. @manickavela29 |
I can try to get for CPU numbers, but i don't have any high performance CPU, (in between someone can add support for dnnl ep 🙂) But here the focus itself is towards GPU with Cuda Vs Trt, is CPU benchmarking relevant? Code blocks are just performance log which I added for zipformer. Those are not part of the patch |
Hi @csukuangfj will create configs for execution provider all together and integrate it with sessions. |
TODO
See https://onnxruntime.ai/docs/execution-providers/TensorRT-ExecutionProvider.html
The text was updated successfully, but these errors were encountered: