-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Details of out-of-distribution #7
Comments
Hi, thank you for your interest in our work. For OOD experiments, M=4 and 16 shots were used. The weights you mentioned, M=1 and 100 shots, were used for zero-shot experiments. We have not released the weights of models used for OOD experiments. If you need those, we could try to find them but not sure if they are still available since the file system of our server recently crashed. Best wishes, |
Hi, Thank you for your response. What is the difference between zero-shot experiments and OOD experiments since both of them trained on Imagenet and tested on downstream tasks. I would appreciate it if you could provide the weights for the OOD experiments. If those are not available, the batch and epoch would also be helpful. I can reproduce it myself. Thanks! |
Hi, they differ in this work regarding the data distribution gap between training and test data. OOD involves the distribution shift only in the input space so the training and test datasets have the consistent category system, while zero-shot involves the distribution shift in both input space and output category so the test input data and/or category may never be seen in the training data. The detailed experimental setups are described in the appendix. Please let me know if there are something missing in the appendix. Lin |
So the cross datasets experiments is the zero-shot experiments right? Thanks for your reply! |
yes, we follow the zero-shot setup in this work: https://arxiv.org/abs/2212.07016. Kindly please close the issue if you feel it is resolved, otherwise, I am happy to answer any other question you have. |
Thank you for sharing your work and code!
In the paper, you claim to use M= 4 and 16 shots when testing the generalization of prompts.
But in the weights you give, it is M= 1 and 100 shots, could you please clarify the detailed experimental setting of out-of-distribution (such as M, shots, batch, and epoch)?
Thanks
The text was updated successfully, but these errors were encountered: