You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello, this is not a bug report, more like asking for advice.
I have an onnx model that's used in a tracking algorithm. But the model originally doesn't accept multiple batches, if there are 4 objects tracked, the model processes the frame 4 times, instead of sending a tensor with a batch of 4.
My question is, is it possible to convert that onnx model to tensorrt, but to make it accept multiple batches at once? so for example if an input should be 1x3x224x224, I would send in 4x3x224x224, so perhaps I could infer all 4 frames with a single call to the engine?
Thank you
The text was updated successfully, but these errors were encountered:
I understand, that to some of you this might sound like a stupid question, but I have had so many problems just convert a model to onnx, then to tensorrt and make it work, my head hurts. I have to dig deep again just to find an answer for this, perhaps anybody knows how to do it?
While some models won't work, e.g. if your model has a reshape node that do the reshape to a fixed shape, then you cannot change the input batch/shape, otherwise it would fail when building engine.
@zerollzeng I found out that the model can be modified in a few places, I managed to make it be able to take and return multiple batches. But now I'm facing a different problem, when adding batches to a tensorrt engine the engines inference time slows down significantly... #3646
Description
Hello, this is not a bug report, more like asking for advice.
I have an onnx model that's used in a tracking algorithm. But the model originally doesn't accept multiple batches, if there are 4 objects tracked, the model processes the frame 4 times, instead of sending a tensor with a batch of 4.
My question is, is it possible to convert that onnx model to tensorrt, but to make it accept multiple batches at once? so for example if an input should be 1x3x224x224, I would send in 4x3x224x224, so perhaps I could infer all 4 frames with a single call to the engine?
Thank you
The text was updated successfully, but these errors were encountered: