Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

GPU memory consumption too high? #1

Closed
jenspetersen opened this issue Dec 13, 2018 · 2 comments
Closed

GPU memory consumption too high? #1

jenspetersen opened this issue Dec 13, 2018 · 2 comments

Comments

@jenspetersen
Copy link

Hey, thanks for implementing GQN in PyTorch! I can only fit batches of size 8 on my Titan Xp (12GB), is that the same for you, or can you fit the default size of 36? In my own implementation I can manage a batch size of 36, but the results don't look too good, so I wanted to try your version :D

Best,
Jens

@iShohei220
Copy link
Owner

iShohei220 commented Dec 13, 2018

Hi!

As you said, this implementation needs very very high computational power because of enormous number of parameters.
This implementation is based on the advice from Dr. Ali Eslami, the first author of GQN paper, so the setting of hyperparameters is absolutely the same as the one on the original paper.
If you have only limited GPU memory, I recommend to use the option of --shared_core True (default: False) or --layers 8 (default: 12) in train.py in order to reduce the number of parameters.
Although the setting would be different with the original paper, you can get enough results as far as I have experimented.

Thank you.

@jenspetersen
Copy link
Author

Hi, yes, I totally missed that this uses separate cores by default. Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants