Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Learning Rate Decay Method #1

Open
kuan-wang opened this issue Mar 9, 2018 · 5 comments
Open

Learning Rate Decay Method #1

kuan-wang opened this issue Mar 9, 2018 · 5 comments

Comments

@kuan-wang
Copy link

Hello, That's really a good result :) Can you share your training method to train mobilenetv2 that high?( such as default learning rate and weight-decay method) . Thanks~
I am currently training the model with SGD and keeping other hyper-parameters the same (except that I use batch size 256), but only get Top1: 71.17

@ericsun99
Copy link
Owner

I have training with SGD, and have using some data augment method.

@SophieZhou
Copy link

@ericsun99 which data augmentation did you use during your training? Using crop, flip, resize, I only got about top 1 accuracy 68 % and top5 about 88%.
And I evaluate the model by center crop.

@SophieZhou
Copy link

@ericsun99 @THUKey In your test, do you test the model with tencrops?

@Coderx7
Copy link

Coderx7 commented Aug 17, 2018

@THUKey ,@ericsun99 could you please share your hyperparameters?

@CF2220160244
Copy link

Hello @THUKey . I use SGD train with a 1080Ti, batchsize=96, lr=0.045, weight decay=0.00004, and decrease the lr 0.98 for each epoch, after 2 days only get 67%, can you tell me your GPU, batchsize, and the initial lr to help me.
I am a student in beijing institute of technology. Thank you very much!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants