-
Notifications
You must be signed in to change notification settings - Fork 6.8k
[Discussion] 1.6.0 Roadmap #15589
Comments
Hey, this is the MXNet Label Bot. |
|
There are a few regressions that have been introduced since the large tensor support was added. For example, there are a few hierarchical attention networks which were training fine until 1.3.1 but since the addition of large tensor support, results in NaNs in weight vectors and gradient calculations. We probably should ensure all regressions are fixed with regard to this feature. Related issues and PRs that might be relevant( the above issue is in addition to what is discussed in the below links) - |
@sandeep-krishnamurthy Agree INT64 enhancements. |
CPU related proposal for 1.6 (keep updating in the following several days).
|
Agree INT64 enhancements, too. The following features are useful for research and development:
|
I think make (pre-trained) NN editable directly may helpful.
Since it is very inconvenient to change the params in ParameterDict(I only knows that use mx.init.Constant may help, but that is too inconvenient for me.), add a What's more, I think MXNet needs a relocatable .dll/.so file I asked how to decrease the dll size (since there is too much The reply is to use
Make the |
It would be nice to add support for default checkpointing in Estimator API. For reference, TF estimator APIs does provide default checkpointing of models trained with Estimator API, which is very useful for users. Also, can we plan to graduate Estimator API in MXNet 1.6? This is super useful API for MXNet users. @roywei - Any comments? |
Thanks to @ChaiBapchya we now have performance comparison data between int32 and int64: https://docs.google.com/spreadsheets/d/1GpdNquQb71Is5B-li99JDuiLeEZd-eSjHIIowzGrwxc/edit#gid=843443107 |
We have multiple improvements to BERT inference and training speed that we would like to be part of 1.6 release:
|
Moving fixing nightly failure from 1.5.1 scope to 1.6.0 as they are failing on master branch not 1.5.x branch.
|
@reminisce @haojin2 given that numpy operators would be a major topic in 1.6 release, it would be great if you could add what you intend to include in this release. |
If I want to use mxnet.gluon.nn.Conv2D to get a depthwise conv layer, I need to explicit set the group argument. Can this be infered automatically? |
As for NumPy compatibility, we would like to add
|
TRT is now working from the CPP package. I think to consider it a released feature we'd want to update documentation and possible target the new TRT version (6). |
I am working on an interface for multi threaded inference in MXNet and it would be great if it could go in 1.6. |
@anirudh2290 this sounds like a larger change. Would you link to the RFC for it? |
@szha yes I am planning to add a RFC this week. |
For reference, users are confused that Large Tensor Support was enabled in MXNet 1.4 and then disabled again in 1.5. Reference: dmlc/gluon-nlp#981 |
Hi I was doing some testing for mxnet 1.6.x and 1.5.1 and I noticed some performance issues in training you can find more details here: #16845 |
Let's start a discussion here about the roadmap towards 1.6.0. We are looking for:
New features that are useful to your research and development.
Improvements and patches to existing features.
If you have any item that you'd like to propose to have in the roadmap, please do:
Create (or locate existing) issue/pull request for the item, note the issue/pull request number.
Comment in this issue: 1) the above issue number, 2) one sentence of what the item is about and why it's useful to you.
Indicate whether you'd be willing to help out on the item.
Share the ETA if you're driving the item and have an guesstimate on when it will be done.
Feel free to include items that weren't included in past roadmap discussions that you still wish to include in this release.
cc @apache/mxnet-committers
The text was updated successfully, but these errors were encountered: