Skip to content

Using Pix2Pix and CycleGAN models on maps dataset and comparing them with FID metric

Notifications You must be signed in to change notification settings

mina-naseh/conditional-gans

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Image-to-Image Translation with Conditional GANs

Hesam Korki, Mina Naseh

Metrics to Evaluate GANs

  • We wanted to know how our changes affect the quality of the generated images
  • GANs are adversarial networks i.e. they use a special type of loss function that both generator and discriminator try to optimize
  • However, the final quality of our output only depends on the generator. So evaluation is not straightforward
  • There are 24 ways of evaluating GANs according to this paper

Fréchet Inception Distance (FID)

  • Introduced in 2017 and still widely adopted
  • It has been shown that FID is consistent with human judgments and is more robust to noise than IS (Inception Score)
  • Generated samples move into a feature space with Inception Net (or any CNN). Viewing the embeddings as a continuous multivariate Gaussian, the mean and covariance are estimated for both the generated data and the real data. Then FID is the distance between the real distribution and the generated one

Step Comparison in pix2pix model

10k 500

We can see that letting the network train for more steps generates more realistic results (as expected) and the FID [Lower better] attests to this.

CycleGAN

We also used CycleGAN to train on maps dataset, and we got a better FID with this method. CycleGAN is constructed based on Pix2Pix. It uses an additional loss function named cycle consistency loss which makes it able to train without paired data. 12epoch

About

Using Pix2Pix and CycleGAN models on maps dataset and comparing them with FID metric

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published