Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Difference from LDM #35

Open
Tsingularity opened this issue Jun 13, 2023 · 1 comment
Open

Difference from LDM #35

Tsingularity opened this issue Jun 13, 2023 · 1 comment

Comments

@Tsingularity
Copy link

Hi, thanks for the great work!

I just noticed that your paper is actually a concurrent work with LDM (exactly the same conference publication!), just wondering what's the main difference between these two works in terms of method? (I took a quick pass but seems that these two papers proposed basically the same technique?)

Thanks!

@createrfang
Copy link

createrfang commented Jun 13, 2023

I am currently learning about both works as well. They are indeed quite similar, but there are also differences. You may notice that LDM (Latent Diffusion) employs a cross-attention mechanism in the context of UNet, whereas Ada-IN (Adaptive Instance Normalization) is used here.

Additionally, VQ-diffusion appears to modify the forward diffusion process and inherits a self-autoregressive mechanism similar to PixelCNN from VQVAE.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants