-
Notifications
You must be signed in to change notification settings - Fork 23
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Learning Rate Decay Method #1
Comments
I have training with SGD, and have using some data augment method. |
@ericsun99 which data augmentation did you use during your training? Using crop, flip, resize, I only got about top 1 accuracy 68 % and top5 about 88%. |
@ericsun99 @THUKey In your test, do you test the model with tencrops? |
@THUKey ,@ericsun99 could you please share your hyperparameters? |
Hello @THUKey . I use SGD train with a 1080Ti, batchsize=96, lr=0.045, weight decay=0.00004, and decrease the lr 0.98 for each epoch, after 2 days only get 67%, can you tell me your GPU, batchsize, and the initial lr to help me. |
Hello, That's really a good result :) Can you share your training method to train mobilenetv2 that high?( such as default learning rate and weight-decay method) . Thanks~
I am currently training the model with SGD and keeping other hyper-parameters the same (except that I use batch size 256), but only get Top1: 71.17
The text was updated successfully, but these errors were encountered: