Q: How do current generative models behave when their (training+sampling) process is iterated? Eg: Train on 50K CIFAR --> Generate 50K samples --> Train on this --> etc. (I imagine this process will amplify deviations... how quickly and to what do things degrade?)
-
-
Very interesting perspective on GAN. Just curious, how expensive CGD would be in GAN training since it seems to require gradient descent on the Hessians? If the purpose is to prevent discriminator from overfitting, wouldn't there be computationally less expensive ways?
-
Current Pytorch implementation requires 7-9 times more gradient calls, depending on network structures and datasets. It is computationally expensive but in principle, it can further speed up if Pytorch can support both forward and backward mode auto differentiation.
End of conversation
New conversation -
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.