-
Notifications
You must be signed in to change notification settings - Fork 3.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Benchmark scripts for training time performance #139
Comments
Yes, you are right! Training speed performance increased once more with this PR. However, training speed may still vary for same GPUs for different PyTorch or CUDA versions. I will try to keep the performance table as up to date as possible and provide training time evaluation scripts for verification. |
Thank you for your respond. I checked my code in message_passing.py. The code is still old and torch embedding is not used. And my environment is ubuntu 16.04, torch 1.0, cuda 9.0 and cudnn7.0. For dataset Cora and Citeseer the experiment results are similar with yours. Only PubMed is different. I only modify dataset name in gcn.py/gat.py. Do I need to modify any other code? |
I added a small script in |
Thank you for your respond. Another question is about distributed computation. In example the dataset is small, for real case the dataset may be large for one worker, so how to separate graph to sub-graph may be a problem, do you have any suggestion? Or dose PyG have the potential to support distributed computation? |
Hi, you can always use more workers if this proves to be beneficial. We support distributed training via |
I mean if the graph is too large for one GPU to store, how to store it? |
Sadly, we currently do not support giant graph processing. Giant graphs are usually processed via sampling techniques. This is a rather difficult but important feature for PyG and it is definitively on my ToDo list. I will close this request in favour of #64. |
❓ Questions & Help
Hi,
I run the gcn.py in /example folder, and change dataset name to "PubMed". In the website the training time about this dataset with gcn.py is 2.0s. But, in my server it only need about 0.7+s. The training time of gat.py is about 3s not 12s. My GPU is GTX 1080 Ti, 200 epochs. So do you know the reason?
The text was updated successfully, but these errors were encountered: