Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Ampere incompatibility? #32

Open
avermilov opened this issue May 22, 2022 · 3 comments
Open

Ampere incompatibility? #32

avermilov opened this issue May 22, 2022 · 3 comments

Comments

@avermilov
Copy link

Hello and thank you so much for the amazing project.

My problem is that the setup process described in the repo seems to not work for Ampere GPUs (in my case RTX 3080 Ti).

First I use the e4e/environment/e4e_env.yaml to create the Conda env. Then I follow the commands in the first cell of stylize.ipynb. However, I get ValueError: Unknown CUDA arch (8.6) or GPU not supported.

I think this may be because the default CUDA installed is 10.x, but my attempts to fix this by setting up the env differently have so far been unsuccessful. Would it be possible to add a fix for Ampere GPUs? Thanks in advance!

@mchong6
Copy link
Owner

mchong6 commented May 22, 2022

The code is written purely in pytorch so this issue seems like incompatibility with pytorch/cuda? Instead of installing using the yaml file, I suggest just installing the pytorch/cuda version that works for you and installing the rest of the required packages.

@avermilov
Copy link
Author

@mchong6 Thank you! Installing a working Pytorch + CUDA and then everything in stylize.ipynb worked perfectly!

Although while the code is working, I seem to not be able to train a 4-image sketch model from scratch because of an immediate CUDA out of memory. I saw you mention that 4-image models are very memory hungry, but do you perhaps know exactly how much memory is needed to train a multi-image model for 2, 4, 8, etc images?

@mchong6
Copy link
Owner

mchong6 commented May 30, 2022

Im not sure how much memory is exactly needed but I could train with 4 images with 20gb memory. If you dont have enough gpu memory, instead of training with a batch size of 4 (1 for each image), you can have a dataloader such that each iteration loads a different image, maintaining batch size of 1.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants