You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The blog post says $500 was spent producing the dataset.
The blog post also says $100 was spent on 3xA100 80GB for 3 hours.
The market rate for 4xA100 is around $8 per hour. (See vast.ai for example)
If the dataset is provided for fine tuning then Alpaca could be reproduce for just about $24 and we would not have to wait for Facebook's response regarding sharing of the pre-trained model.
The text was updated successfully, but these errors were encountered:
MarkSchmidty
changed the title
Reduce reproduction cost from $600 to $24 by releasing the instruct dataset only
Reduce reproduction cost from 96%, from $600 to $24, by releasing the instruct dataset only
Mar 13, 2023
MarkSchmidty
changed the title
Reduce reproduction cost from 96%, from $600 to $24, by releasing the instruct dataset only
Reduce reproduction cost 96%, from $600 to $24, by releasing the instruct dataset only
Mar 13, 2023
You're right alpaca_data.json is our released dataset. We're also releasing the recipe for producing the dataset, so other researchers can build on this.
You're correct that excluding the cost of reproducing the data, the cost of training the model is much lower.
The blog post says $500 was spent producing the dataset.
The blog post also says $100 was spent on 3xA100 80GB for 3 hours.
The market rate for 4xA100 is around $8 per hour. (See vast.ai for example)
If the dataset is provided for fine tuning then Alpaca could be reproduce for just about $24 and we would not have to wait for Facebook's response regarding sharing of the pre-trained model.
The text was updated successfully, but these errors were encountered: