Pix2Pix

Generative Models PyTorch Aerial2Map
GitHub →

Overview

From-scratch replication of Pix2Pix for paired image-to-image translation. Unlike CycleGAN, Pix2Pix requires paired training data and uses a U-Net generator with a PatchGAN discriminator that classifies overlapping 70×70 image patches as real or fake. Based on Image-to-Image Translation with Conditional Adversarial Networks (Isola et al., 2017).

Architecture

Generator: U-Net with skip connections (encoder-decoder + lateral connections at each resolution)

Discriminator: PatchGAN — classifies 70×70 patches rather than the full image, encouraging sharp local texture

  • Combined adversarial + L1 reconstruction loss
  • L1 weight encourages low-frequency correctness

Training

  • Dataset: Aerial2Map (aerial photography ↔ map tiles)
  • Framework: PyTorch

Paper

Image-to-Image Translation with Conditional Adversarial Networks — Isola et al., 2017