Skip to content
This repository has been archived by the owner on Oct 19, 2024. It is now read-only.

[BLOOM] Support encoder chunk sizes > 1 #738

Merged
merged 9 commits into from
Oct 9, 2022

Conversation

ddxxdd-code
Copy link
Contributor

Add support for alpa's internal padding in bloom_model.py
Remove warnings on encoder chunk sizes in wrapper.py

ddxxdd-code and others added 8 commits October 3, 2022 10:45
Add unk_token_id of bloom, set special token_ids to their values in bloom for inference with bloom.
Support alpa's internal padding in attention masks to enable higher encoder chunk sizes
Remove warnings for larger encoder chunk sizes for bloom model as we are supporting them now.
Add support for all encoder chunk sizes
Remove useless if statement as attention mask is always passed to bloom model.
@merrymercy merrymercy force-pushed the main branch 2 times, most recently from b5bdb0d to fe6bd5f Compare October 9, 2022 08:54
@merrymercy merrymercy merged commit fbcb2ab into alpa-projects:main Oct 9, 2022
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants