-
Notifications
You must be signed in to change notification settings - Fork 531
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add Composer MPT to FasterTransformer Conversion Script #519
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
looks good to me, but would appreciate if someone more familiar with the conversion scripts could also take a look!
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can we pull out the shared code here between the two FT conversion scripts and put it in llmfoundry proper? This will also allow us to more easily write a callback that saves these checkpoints during training.
…lm-foundry into nik/conversion-script
…lm-foundry into nik/conversion-script
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
@nik-mosaic want to merge? |
Directly convert a Composer MPT model to the FasterTransformer format, without an intermediate conversion to HuggingFace.
Associated task: CO-2277
The amount of CPU RAM required to run this script is approximately equal to the size of the MPT checkpoint.pt file.
Example:
Command:
python convert_composer_mpt_to_ft.py --composer_path=<folder> --ft_save_dir='ft' --infer_gpu_num=1 --output_precision='fp16'
Output:
<TODO: Add image of output comparison>