Skip to content

Latest commit

 

History

History
10 lines (8 loc) · 393 Bytes

README.md

File metadata and controls

10 lines (8 loc) · 393 Bytes

Pay-Attention-to-MLPs

Implementation of the gMLP model introduced in Pay Attention to MLPs.

The authors of the paper propose a simple attention-free network architecture, gMLP, based solely on MLPs with gating, and show that it can perform as well as Transformers in key language and vision applications.