Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can you open source the ImageNet Pretraining code? #32

Open
shiyutang opened this issue Dec 13, 2022 · 7 comments
Open

Can you open source the ImageNet Pretraining code? #32

shiyutang opened this issue Dec 13, 2022 · 7 comments

Comments

@shiyutang
Copy link

I have tried to pre-training the backbone+ppa+sase on imagenet, with convbnconv + linear as head. But I cannot reproduce the result of 75.3 on base model. So can you provide the source code of pertaining on imagenet?

@speedinghzl
Copy link
Member

Please refer to https://github.com/fudan-zvg/SeaFormer for the ImageNet training script.

@shiyutang
Copy link
Author

Thank you a lot. I will try it. Did you do anything in particular on imagenet pertaining? Since I cannot reproduce the imagenet top1 ACC. I manage to distill the model to achieve the ACC goal, but the training mIOU on ADE20k is still unsatisfactory.

@speedinghzl
Copy link
Member

Maybe the training hyperparameters do matter. The training details are listed here,
https://github.com/fudan-zvg/SeaFormer/tree/main/seaformer-cls#training.

@shiyutang
Copy link
Author

Thank you very much :) The training process is undergoing. I have read the seaformer and thought it is very amazing. Good work!

@shiyutang
Copy link
Author

After I train on seaformer-cls framework, the Imagenet top 1 acc goes from 72.6 to 74.9(Original implementation is 75.3). But the segmentation mIOU on ADE20K is almost the same as 36.68(It should be near 39.0), what could be the thing that I am missing?

@speedinghzl
Copy link
Member

speedinghzl commented Feb 7, 2023

Thanks for the update.
It seems that you have reproduced the classification accuracy. Did you meet NAN or something unnormal when training on Imagenet? After training on ImageNet, there is nothing special about the training segmentation model.

Do you use the official Topformer segmentation code or MMsegmentation? There is a small difference between them.
Besides, Do you reproduce the 39.2 mIoU with the provided ImageNet pre-trained model?

@shiyutang
Copy link
Author

I have reproduced the 38.3 mIOU(bs=16) with the provided model using the official topformer model. And there is nothing strange when training the imagenet.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants