Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about the training hyperparameters (epoch/lr) #1

Closed
skylineo opened this issue Dec 13, 2024 · 2 comments
Closed

Question about the training hyperparameters (epoch/lr) #1

skylineo opened this issue Dec 13, 2024 · 2 comments

Comments

@skylineo
Copy link

Thank you for your work, this is an excellent piece of code. I’ve recently been trying to apply different pretraining strategies to trajectory prediction networks. In your code, I noticed that the pretraining epoch is set to 20, the fine-tuning epoch is set to 40, and the learning rate is the same for both (ignoring the differences due to cosine decay). However, in other papers, I’ve found that in some cases, the pretraining epochs and learning rates are much larger than for fine-tuning (e.g., SEPT, TrajMAE), or both parts use the same hyperparameters (e.g., Forecast-MAE). Currently, I’m planning to modify your code. Could you advise on how to set the pretraining and fine-tuning epochs and learning rates more reasonably?Is 20 epochs for pretraining and 40 epochs for fine-tuning a reasonable choice?

@KexianShen
Copy link
Owner

I didn't think much about the epoch and learning rate at the time, as the pre-training results were quite good. Your suggestion is excellent.

Wishing you better results in your experiments.

@skylineo
Copy link
Author

Got it! Thanks for your timely response!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants