Skip to content

Latest commit

 

History

History
16 lines (14 loc) · 1.08 KB

dl_22.md

File metadata and controls

16 lines (14 loc) · 1.08 KB

Unpaired image-to-image translation using cycle-consistent adversarial networks

Jun-Yan Zhu, Taesung Park, Phillip Isola, Alexei A. Efros (2017)

Key points

  • Learning image-to-image translation without need for paired examples
    • Capture special characteristics of one image collection and figure out how these could be translated into the other collection
  • Cycle-consistent: translation of translation should be same as original --> ensure input-output is paired in a meaningful way!
    • 2 generative adversarial networks (GANs), 2 consistency losses
    • Only 1 GAN and no cycle-consistency: mode collapse (all input --> same output) or irrelevant mappings
  • GAN: discriminator "helps" generator through backprop (gives info on how to improve)
    • Endpoint of training: discriminator no better than random guessing (0.5)

Limitations

  • Only works for color/texture changes, not for geometric ones
  • Other variations, not present in training data: e.g. network thinks man on horse is part of the horse
  • Weak semantic supervision may be required to close the gap with paired translators