Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Example/Test Model Benchmarks (Canonical WandB runs) #74

Closed
cat-state opened this issue Oct 27, 2022 · 6 comments
Closed

Example/Test Model Benchmarks (Canonical WandB runs) #74

cat-state opened this issue Oct 27, 2022 · 6 comments
Labels
feature request New feature or request

Comments

@cat-state
Copy link
Collaborator

🚀 The feature, motivation, and pitch

If we had links to benchmarks for the example (and/or test) models, it would be easier to add new models, and keep track of improvements in method implementations. Additionally, during refactoring, it would allow checking that no performance degrading changes were introduced.

This can be a minimal version of #13

Alternatives

No response

Additional context

No response

@albertsun1
Copy link

Hey! I'm new to contributing to trlx, would it be worth for me to give this a go for the ppo/ilql sentiment examples?

@maxreciprocate
Copy link
Collaborator

@cat-state something like that? vwxyzjn/cleanrl#307

@cat-state
Copy link
Collaborator Author

cat-state commented Nov 11, 2022

@albertsun1

Hey! I'm new to contributing to trlx, would it be worth for me to give this a go for the ppo/ilql sentiment examples?

Sure, although you might need compute?
@reciprocated maybe we should make a single-node config version that can be finetuned on a single gpu fast?

@cat-state
Copy link
Collaborator Author

cat-state commented Nov 11, 2022

So I see that WandB actually lists the commit hash used for a run. So if we could find/tag TRLX runs in wandb then each commit could be matched up to a specific state of the repository.

@maxreciprocate
Copy link
Collaborator

maxreciprocate commented Nov 11, 2022

Sure, although you might need compute? @reciprocated maybe we should make a single-node config version that can be finetuned on a single gpu fast?

fwiw {ppo,ilql}_config.yml were meant to be single gpu, up to a batch_size, since they both use gpt2 small

@cat-state cat-state added the feature request New feature or request label Feb 2, 2023
@maxreciprocate
Copy link
Collaborator

Resolved with #357

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature request New feature or request
Projects
None yet
Development

No branches or pull requests

3 participants