We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hi, authors:
Great work! While, I'm a bit confused about the description and code, in paper A.5: x_t is called noisy image. However, in code,
if self.cfg.recon_loss: # reconstruct x0 latents_recon = self.model.predict_start_from_noise( latents_noisy, t, noise_pred ) # x0-reconstruction loss from Sec 3.2 and Appendix loss = ( 0.5 * F.mse_loss(latents, latents_recon.detach(), reduction="sum") / latents.shape[0] ) grad = torch.autograd.grad(loss, latents, retain_graph=True)[0]
x_0, x_t is actually latent after vae and noisy latent, if correct.
There do exist methods that apply loss on image-space such as HiFA, and ReconFusion, which may be confusing.
Please clearify that I'm understanding it right, Thanks!
The text was updated successfully, but these errors were encountered:
No branches or pull requests
Hi, authors:
Great work! While, I'm a bit confused about the description and code, in paper A.5:

x_t is called noisy image. However, in code,
x_0, x_t is actually latent after vae and noisy latent, if correct.
There do exist methods that apply loss on image-space such as HiFA, and ReconFusion, which may be confusing.
Please clearify that I'm understanding it right, Thanks!
The text was updated successfully, but these errors were encountered: