Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Configuration to reproduce the cifar-10 ct adaptive training result in the paper. #4

Open
Zyriix opened this issue Oct 7, 2023 · 1 comment

Comments

@Zyriix
Copy link

Zyriix commented Oct 7, 2023

Thanks for your innovative work!

I'm new to the consistency model and trying to reproduce the FID curve in your paper(Fig. 3d) with PyTorch. However, when using the configuration in this repo. I found the FID does not converge like Fig. 3d. I think this might be a reason for the following configuration:

  1. warmup: I use no warmup following the paper. I tried to use a 10M warmup like this repo, but the FID converges much slower than Fig 3d.
  2. learning rate(I adopt 4e-4)
  3. ema_weight to generate samples(following paper I use 0.9999)
  4. FIR resampling: I did not implement FIR kernel in my implementation
  5. Fourier Pos Embedding: I use sin/cos pos embedding in my implementation

I'm quite sure I made the right implementation for other components (like adaptive u, adaptive T, and skip factor c).

I reached an FID of 12.84 at CIFAR-10 using clean-fid's pytorch-legacy statistics.

Do you think FIR and Fourier's pos embedding is important for reproducing the result in the paper?
What is the configuration for reproducing the FID curve in Fig. 3d?

Thanks again for your generous contributions!

@Zyriix
Copy link
Author

Zyriix commented Oct 7, 2023

I use the u-net model in your PyTorch repo, and the configuration following this repo, with
use_scale_shift_norm=True
resblock_updown=True
num_head_channels=64

@Zyriix Zyriix changed the title Configuration to reproduce the cifar-10 ct adaptive training result in paper. Configuration to reproduce the cifar-10 ct adaptive training result in the paper. Oct 7, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant