You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In your CGAN implementation, you feed G and D with the same real label from the dataset as the condition.However, I have a question about this since several other implementations feed G with a sampled label as the condition.
I have seen several CGAN implementations. There are two ways to use the condition labels:
feed G with a random label, feed D with the real label
As i know, specific implementation is not referred in the cGAN paper. But the Figure 1 at page 3, y vector is used as the same manner. so, i guess the author maybe used the same condition label for both of them.
actually, i think there's no need to be trained with the same condition label for both G and D. In the aspect of the implementing the code, using the same condition label is easier than the different condition label (cuz we need to generate another sample for D). and i also think using the same condition label can lead more stable training and is more make sense for the purpose of D (it is not sure 100%).
Hi @kozistr, Thanks for your great work!
In your CGAN implementation, you feed G and D with the same real label from the dataset as the condition.However, I have a question about this since several other implementations feed G with a sampled label as the condition.
I have seen several CGAN implementations. There are two ways to use the condition labels:
Here are my questions:
The same issue has been asked here. I hope we can figure out it together.
The text was updated successfully, but these errors were encountered: