Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Change optimizer or lr_scheduler in resuming training without removing the global_step information #20552

Open
arijit-hub opened this issue Jan 17, 2025 · 0 comments
Labels
feature Is an improvement or enhancement needs triage Waiting to be triaged by maintainers

Comments

@arijit-hub
Copy link

arijit-hub commented Jan 17, 2025

Description & Motivation

Hello,

In case we want to resume training from a checkpoint, but for say change the optimizer class or the lr_scheduler class, it seems the global_step becomes 0. Is there a possibility to keep the global step information and still allow the change? This would greatly benefit if in case we are using a warmup + decay lr_scheduler, and on resume training, we want to change the number of decay steps. However with this change if the global step becomes 0, the model will do a warmup at resume, which is not intended.

Pitch

No response

Alternatives

No response

Additional context

No response

cc @lantiga @Borda

@arijit-hub arijit-hub added feature Is an improvement or enhancement needs triage Waiting to be triaged by maintainers labels Jan 17, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature Is an improvement or enhancement needs triage Waiting to be triaged by maintainers
Projects
None yet
Development

No branches or pull requests

1 participant