# GAN Papers with 1k+ citations
###### tags: `GAN`
Updated on Feb. 2021
1. (Vanilla GAN)[Goodfellow, I. J., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., ... & Bengio, Y. (2014). Generative adversarial networks. arXiv preprint arXiv:1406.2661.](https://arxiv.org/abs/1406.2661) Cited by 27497, 1st GAN paper
2. (DCGAN)[Radford, A., Metz, L., & Chintala, S. (2015). Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434.](https://arxiv.org/abs/1511.06434) Cited by 8447, improve CNN in GAN using striding
3. (Pix2Pix)[Isola, P., Zhu, J. Y., Zhou, T., & Efros, A. A. (2017). Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1125-1134).](https://arxiv.org/abs/1611.07004v3) Cited by 7899, Nvidia change noise-to-picture to picture-to-picture
4. (cycleGAN)[Zhu, J. Y., Park, T., Isola, P., & Efros, A. A. (2017). Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE international conference on computer vision (pp. 2223-2232).](https://arxiv.org/abs/1703.10593v6) Cited by 7325, most well-known cycle loss with pix2pix discrimination
5. (WGAN)[Arjovsky, M., Chintala, S., & Bottou, L. (2017, July). Wasserstein generative adversarial networks. In International conference on machine learning (pp. 214-223). PMLR.](https://arxiv.org/abs/1701.07875) Cited by 6202, replace JS divergence with Wasserstein distance
6. (SRGAN)[Ledig, C., Theis, L., Huszár, F., Caballero, J., Cunningham, A., Acosta, A., ... & Shi, W. (2017). Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 4681-4690).](https://arxiv.org/abs/1609.04802v5) Cited by 5026, the most well-known super-resolution paper
7. (conditional GAN)[Mirza, M., & Osindero, S. (2014). Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784.](https://arxiv.org/abs/1411.1784) Cited by 4729, 1st conditional GAN
8. (WGAN-GP)[Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., & Courville, A. (2017). Improved training of wasserstein gans. arXiv preprint arXiv:1704.00028.](https://arxiv.org/abs/1704.00028)Cited by 4187, accelerated WGAN
9. (DANN)[Ganin, Y., Ustinova, E., Ajakan, H., Germain, P., Larochelle, H., Laviolette, F., ... & Lempitsky, V. (2016). Domain-adversarial training of neural networks. The journal of machine learning research, 17(1), 2096-2030.](https://www.jmlr.org/papers/volume17/15-239/15-239.pdf), Cited by 2737 Domain Adaptation
10. (ProGAN)[Karras, T., Aila, T., Laine, S., & Lehtinen, J. (2017). Progressive growing of gans for improved quality, stability, and variation. arXiv preprint arXiv:1710.10196.](https://arxiv.org/abs/1710.10196), Cited by 2597, layer by layer learning improve quality
11. (InfoGAN)[Chen, X., Duan, Y., Houthooft, R., Schulman, J., Sutskever, I., & Abbeel, P. (2016). Infogan: Interpretable representation learning by information maximizing generative adversarial nets. arXiv preprint arXiv:1606.03657.](https://arxiv.org/abs/1606.03657) Cited by 2522, adding conditional variables into input noise for conditional GAN
12. (2time scale update)[Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., & Hochreiter, S. (2017). Gans trained by a two time-scale update rule converge to a local nash equilibrium. arXiv preprint arXiv:1706.08500.](https://arxiv.org/abs/1706.08500) Cited by 2235, run discriminator more times to improve training
13. (LSGAN)[Mao, X., Li, Q., Xie, H., Lau, R. Y., Wang, Z., & Paul Smolley, S. (2017). Least squares generative adversarial networks. In Proceedings of the IEEE international conference on computer vision (pp. 2794-2802).](https://arxiv.org/abs/1611.04076) Cited by 1991, replace Wasserstein distance with JS divergence and L2 error of a linear output
14. (ADDA)[Tzeng, E., Hoffman, J., Saenko, K., & Darrell, T. (2017). Adversarial discriminative domain adaptation. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 7167-7176).](https://arxiv.org/abs/1702.05464) Cited by 1924, apply GAN on classifier domain transfer
15. (GAN-CLS)[Reed, S., Akata, Z., Yan, X., Logeswaran, L., Schiele, B., & Lee, H. (2016, June). Generative adversarial text to image synthesis. In International Conference on Machine Learning (pp. 1060-1069). PMLR.](https://arxiv.org/abs/1605.05396)Cited by 1860, GAN text to image synthesis
16. (LAPGAN)[Denton, E., Chintala, S., Szlam, A., & Fergus, R. (2015). Deep generative image models using a laplacian pyramid of adversarial networks. arXiv preprint arXiv:1506.05751.](https://arxiv.org/abs/1506.05751) Cited by 1791, learning self upsample
17. (Spectral Normalization)[Miyato, T., Kataoka, T., Koyama, M., & Yoshida, Y. (2018). Spectral normalization for generative adversarial networks. arXiv preprint arXiv:1802.05957.](https://arxiv.org/abs/1802.05957) Cited by 1750, real 1-Lipschitz GAN loss
18. (styleGAN)[Karras, T., Laine, S., & Aila, T. (2019). A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 4401-4410).](https://openaccess.thecvf.com/content_CVPR_2019/html/Karras_A_Style-Based_Generator_Architecture_for_Generative_Adversarial_Networks_CVPR_2019_paper.html) Cited by 1634, AdaIN style transfer with high quality
19. (stackGAN)[Zhang, H., Xu, T., Li, H., Zhang, S., Wang, X., Huang, X., & Metaxas, D. N. (2017). Stackgan: Text to photo-realistic image synthesis with stacked generative adversarial networks. In Proceedings of the IEEE international conference on computer vision (pp. 5907-5915).](https://openaccess.thecvf.com/content_iccv_2017/html/Zhang_StackGAN_Text_to_ICCV_2017_paper.html)Cited by 1475, text to image 2 stage image improvement
20. (BigGAN)[Brock, A., Donahue, J., & Simonyan, K. (2018). Large scale GAN training for high fidelity natural image synthesis. arXiv preprint arXiv:1809.11096.](https://arxiv.org/abs/1809.11096) Cited by 1455, stabilize training with large batch + truncated sampling input noise
21. (UNIT)[Liu, M. Y., Breuel, T., & Kautz, J. (2017). Unsupervised image-to-image translation networks. arXiv preprint arXiv:1703.00848.](https://arxiv.org/abs/1703.00848)Cited by 1441, common latent space for domain A and B
22. (StarGAN)[Choi, Y., Choi, M., Kim, M., Ha, J. W., Kim, S., & Choo, J. (2018). Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 8789-8797).](https://openaccess.thecvf.com/content_cvpr_2018/html/Choi_StarGAN_Unified_Generative_CVPR_2018_paper.html)Cited by 1398, multiple cross-domain style transfer
23. (MIXUP)[Zhang, H., Cisse, M., Dauphin, Y. N., & Lopez-Paz, D. (2017). mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412.](https://arxiv.org/abs/1710.09412) Cited by 1389, change Empirical Risk Minimization to Vincinal Risk Minimization for data augmentation
24. (SeqGAN)[Yu, L., Zhang, W., Wang, J., & Yu, Y. (2017, February). Seqgan: Sequence generative adversarial nets with policy gradient. In Proceedings of the AAAI conference on artificial intelligence (Vol. 31, No. 1).](https://ojs.aaai.org/index.php/AAAI/article/view/10804)Cited by 1330, RL sequence to sequence GAN
25. (SAGANs)[Zhang, H., Goodfellow, I., Metaxas, D., & Odena, A. (2019, May). Self-attention generative adversarial networks. In International conference on machine learning (pp. 7354-7363). PMLR.](https://arxiv.org/abs/1805.08318)Cited by 1327, intra-image self-attention GAN
26. (DiscoGAN)[Kim, T., Cha, M., Kim, H., Lee, J. K., & Kim, J. (2017, July). Learning to discover cross-domain relations with generative adversarial networks. In International Conference on Machine Learning (pp. 1857-1865). PMLR.](http://proceedings.mlr.press/v70/kim17a.html)Cited by 1180, cycle GAN with single discrimination
27. (Problem of JS divergence)[Arjovsky, M., & Bottou, L. (2017). Towards principled methods for training generative adversarial networks. arXiv preprint arXiv:1701.04862.](https://arxiv.org/abs/1701.04862)Cited by 1140, JS divergence cannot distinguish between those untouched conditions
28. (GAIL)[Ho, J., & Ermon, S. (2016). Generative adversarial imitation learning. arXiv preprint arXiv:1606.03476.](https://arxiv.org/abs/1606.03476)Cited by 1083, GAN in imitation learning in RL
29. (3D-GAN)[Wu, J., Zhang, C., Xue, T., Freeman, W. T., & Tenenbaum, J. B. (2016). Learning a probabilistic latent space of object shapes via 3d generative-adversarial modeling. arXiv preprint arXiv:1610.07584.](https://arxiv.org/abs/1610.07584)Cited by 1063, GAN 3d object
30. (CyCADA)[Hoffman, J., Tzeng, E., Park, T., Zhu, J. Y., Isola, P., Saenko, K., ... & Darrell, T. (2018, July). Cycada: Cycle-consistent adversarial domain adaptation. In International conference on machine learning (pp. 1989-1998). PMLR.](https://arxiv.org/abs/1711.03213) Cited by 1066, cycle GAN in domain adaptation
31. (DualGAN)[Yi, Z., Zhang, H., Tan, P., & Gong, M. (2017). Dualgan: Unsupervised dual learning for image-to-image translation. In Proceedings of the IEEE international conference on computer vision (pp. 2849-2857).](https://arxiv.org/abs/1704.02510) Cited by 1014, cycle GAN with U-Net
## Type of Papers (order by publish time)
### Origian GAN type
2014 Vanilla GAN, 27497 cites
2014 conditional GAN, 4729 cites
2015 DCGAN, 8447 cites
### Unpaired GAN
2017 cycleGAN, 7325 cites
2017 UNIT, 1441 cites
2017 DualGAN, 1014 cites
2017 DiscoGAN, 1180 cites
2018 StarGAN, 1398 cites
### Higher resolution
2015 LAPGAN (generated resolution), 1791 cites
2016 SRGAN (upsamled resolution), 5026 cites
2017 ProGAN (generated resolution), 2597 cites
2018 SAGANs (generated resolution), 1327 cites
2019 styleGAN (generated resolution), 1634 cites
### Better Embedding Condition
2016 InfoGAN, 2522 cites
2016 Pix2Pix, 7899 cites
2017 MIXUP, 1389 cites
2018 BigGAN, 1455 cites
### Handeling optimization issues
2016 LSGAN, 1991 cites
2017 Problem of JS divergence, 1140 cites
2017 WGAN, 6202 cites
2017 WGAN-GP, 4187 cites
2017 2time scale update, 2235 cits
2018 Spectral Normalization, 1750 cites
### Domain Adaptation
2016 DANN, 2737 cites
2017 ADDA, 1924 cites
2017 CyCADA, 1066 cites
### Text to image
2016 GAN-CLS, 1860 cites
2017 stackGAN, 1475 cites
### Text to Text
2017 SeqGAN, 1330 cites
### RL
2016 GAIL, 1083 cites
### 3D object synthesis
2016 3D-GAN, 1063 cites