Implementation of the paper: "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"
-
Notifications
You must be signed in to change notification settings - Fork 0
Implementation of the paper: "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"
License
eclouder/MoE-LLM
About
Implementation of the paper: "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"
Resources
License
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published