-
Notifications
You must be signed in to change notification settings - Fork 7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
TorchVision Roadmap - 2021 H1 #3221
Comments
Can we revisit various workarounds for tracing Faster RCNN / MaskRCNN ? There are many such special code-path for tracing mode, to support ONNX export. These code were added one year ago when PyTorch ONNX had many limitation. Especially the one below does repeated concat in a loop so it is not efficient. vision/torchvision/models/detection/roi_heads.py Lines 454 to 461 in 3d60f49
Now that PyTorch ONNX export has support for lists and also inplace assignment, I think many of those workarounds can be removed. Even if removing is not possible, we should be able to make them more efficient, thanks to the improved ONNX export support in recent PyTorch. For example, The snippet above can be replaced with the one below, which uses a batched concat. I confirmed that ONNX export of maskrcnn works with this change applied. This is also what TVM prefers.
|
Just linking a few trackers here |
How about adding RandAugment, in addition to already implemented AutoAugment ? |
Hi, I was just wondering if there are any plans to add support for the QNNpack backend to the quantised ResNet50 model? (this would enable quantised ResNet50 for AArch64, which isn't supported by the FBGEMM backend). |
Is the newly released d2go mentioned here? |
Thoughts about FastAutoAugment ? |
Implemented in #3792 |
Rotated Boxes RFC #2761. |
can you please open a new issue so that we can track this?
yes, exactly
Can you open a new issue so that we can track it down and discuss it there? |
Is there any interest in adding MoViNets or some other state of the art models in video understanding? |
Comments and suggestions are welcome :) ! |
@Atze00 there's no current plan to support MoViNets but regarding video understanding, we're planning to support Optical Flow with a RAFT implementation
Could you clarify what you mean? I'm not sure this answers your question but in general, when we implement a model in torchvision we provide the architecture, the pre-trained weights, and also some training receipe |
@NicolasHug My question was based on the fact that I've been working on implementing movinets in pytorch, using the weights released by the authors in TF. I've been wondering if the code alone would be enough to discuss about a possible addition of this architecture into the roadmap. Unfortunately I don't have the resources or the experience necessary to reproduce state of the art results in video understanding. |
@fmassa @NicolasHug Shall we close this issue in favor of #4187? Since only the rotated boxes RFC is still open we could consider moving it on H2 or omit it if nobody wants to pick it up. |
Moved the pending to H2 at #4187, closing this issue. |
This issue is be used to keep track of the roadmap of TorchVision and get feedback from our users.
For the 2021-H2 roadmap, please see #4187
The text was updated successfully, but these errors were encountered: