-
Notifications
You must be signed in to change notification settings - Fork 21
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
the retrieval loss doesn't converge well #11
Comments
thats very common in the first several epochs. Try training it a little bit longer. Or just restart the training. |
ok, thanks a lot, for another VSE model(VSEAttModel) and "pair loss" , whose result isn't shown in your paper "Discriminability objective for training descriptive captions" in CVPR 2018? |
Pair loss is worse and vseattmodel gives worse result too. |
thanks!if the retrieval model perform better(like the paper“Stacked Cross Attention for Image-Text Matching”),can we get a better result for captioning model? |
I think it's very likely. |
hello,luo |
Did you download my pretrained model? Does it perform better and the same as what's reported in the paper? |
i might get the problem,i have used the size of 7x7 for coco fc features, i think u have used 14x14 for coco fc features? |
fc feature doest have spatial dimensions, it's a vector |
I found other paper use Karpathy'split for COCO, your paper use rama's split, whose test data are the same? why you can compare your result with the result in self-critical? |
the splits are different. The self critical one is my implementation on Rama's split. Using Rama split I'd because we need to compare ours to Rama's result. |
Hello, luo
when I pretrain the VSEFCmodel, the vse_loss doesn't converge well , just around 51.2. is there some mistakes in my experiments, how about your vse_loss when you pretrain VSEFCmodel?
The text was updated successfully, but these errors were encountered: