Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions about computational complexity #21

Open
AJDA1992 opened this issue Oct 31, 2022 · 0 comments
Open

Questions about computational complexity #21

AJDA1992 opened this issue Oct 31, 2022 · 0 comments

Comments

@AJDA1992
Copy link

AJDA1992 commented Oct 31, 2022

First I want to say thank you for releasing the code for this excellent work. I have a couple of observations/questions that hopefully you can respond to.

It appears that the reported 645 M Flops is reported on a per frame basis which would imply that for the 243 frame model the flops for a forward paass would be 645*243 M FLOPS is this correct?

I am also witnessing that to train a T=27 model on 2 3090 RTXs (48 GB of VRAM) I can only use a batch size of 200 (this is the max before Memory errors)). Can you please tell us what gpus you trained on to be able to support a batch size of 1024?

I am wondering if the small increases in performance wrt to accuracy are outweighed by the extreme computational overhead which seems to be making training almost intractable for the majority.

@AJDA1992 AJDA1992 changed the title Serious doubts about computational complexity Questions about computational complexity Oct 31, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant