Skip to content

Artificial Neural Network training convergence boosting by smart data ordering

License

Notifications You must be signed in to change notification settings

Tensor-Reloaded/Convergence

Repository files navigation

Convergence

Artificial Neural Network training convergence boosting by smart data ordering.

The project relies on the assumtion that there must be an optimal order in which training data is fed to a model. We assume this order would improve the convergence speed of the training process, effectively requiring less time and computing power.

This optimal order will be approximated using metrics collected dynamicly during training, such as sample/batch loss, gradients, input-ouput distance etc.

This convergence boosting method is promissing since it could be applied to any ANN during training in order to speed it up

I also want to underline that, we can experiment with creating batches based on a distribution and then "epochs" would not be accurate, we would simply have number of batches. In this scenario, it won't be a "ordering" method, but a batch "sampling" method.

About

Artificial Neural Network training convergence boosting by smart data ordering

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •