You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I truly appreciate the FLM-101B team open-sourcing this large-scale language model! After reading the paper, I also have some thoughts on optimizing the training framework, mainly in these aspects:
Progressive data selection strategy: using different datasets for models of different scales to achieve gradual enhancement.
Parameter update driven growth: inserting new layers based on layer update status.
Layer-wise learning rates: setting independent learning rates for different layers.
Genetic algorithm based model expansion.
Incremental fine-tuning for transfer learning.
I drafted a document elaborating these ideas in details. If the team finds it relevant, I'd be very happy to have the opportunity to further discuss thoughts on optimizing the training framework. Please reply to this issue or contact me via [email protected].
Again, thank you for the contributions of the FLM-101B team!
The text was updated successfully, but these errors were encountered:
I truly appreciate the FLM-101B team open-sourcing this large-scale language model! After reading the paper, I also have some thoughts on optimizing the training framework, mainly in these aspects:
Progressive data selection strategy: using different datasets for models of different scales to achieve gradual enhancement.
Parameter update driven growth: inserting new layers based on layer update status.
Layer-wise learning rates: setting independent learning rates for different layers.
Genetic algorithm based model expansion.
Incremental fine-tuning for transfer learning.
I drafted a document elaborating these ideas in details. If the team finds it relevant, I'd be very happy to have the opportunity to further discuss thoughts on optimizing the training framework. Please reply to this issue or contact me via [email protected].
Again, thank you for the contributions of the FLM-101B team!
The text was updated successfully, but these errors were encountered: