새소식

Generative Models

생성 모델 논문 추천 List - Generative Adversarial Networks

  • -

Generative Adversarial Networks(GAN)는 생성 모델의 한 종류입니다. GAN의 등장이후 생성모델은 매우 빠르게 발전해 오고 있으며 개인적으로 생성모델을 공부한다면 필수적으로 공부해야할 부분중 하나라고 생각합니다. Diffusion model 등 최신 트랜드의 논문들도 GAN을 기반으로 하거나 GAN에서 나온 이론을 토대로 하는 경우가 많습니다.

 

따라서 필수적으로 읽어야 하는 GAN 관련 논문들을 정리해보려고 합니다. 사실 논문이 워낙 많기 때문에 빠트린 논문들도 많습니다. (계속해서 업데이트 할 생각입니다.) 논문을 볼때는 블로그 리뷰를 읽는 것도 좋지만, 논문의 핵심적인 부분 외에도 자잘한 insight 들과 Tip들이 큰 도움이 되니 웬만하면 논문의 Method 부분은 꼭 읽는 것을 추천드립니다. 그리고 논문을 읽다가 생소한 부분(Auto Encoder, Vector Quantization 등)이 나오면 넘어가지 말고 블로그 리뷰나 reference에 있는 논문을 찾아서 읽으면 큰 도움이 될것 같습니다.

 

(논문에 대한 설명도 추가할 예정.. 언젠가는..)

 


GAN (2014)

논문 제목: Generative Adversarial Networks

https://arxiv.org/abs/1406.2661

 

Generative Adversarial Networks

We propose a new framework for estimating generative models via an adversarial process, in which we simultaneously train two models: a generative model G that captures the data distribution, and a discriminative model D that estimates the probability that

arxiv.org

 

CGAN (2014)

논문 제목: Conditional Generative Adversarial Nets

https://arxiv.org/abs/1411.1784

 

Conditional Generative Adversarial Nets

Generative Adversarial Nets [8] were recently introduced as a novel way to train generative models. In this work we introduce the conditional version of generative adversarial nets, which can be constructed by simply feeding the data, y, we wish to conditi

arxiv.org

 

DCGAN (2015)

논문 제목: Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks

https://arxiv.org/abs/1511.06434

 

Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks

In recent years, supervised learning with convolutional networks (CNNs) has seen huge adoption in computer vision applications. Comparatively, unsupervised learning with CNNs has received less attention. In this work we hope to help bridge the gap between

arxiv.org

 

Improved GAN (2016)

논문 제목: Improved Techniques for Training GANs

https://arxiv.org/abs/1606.03498

 

Improved Techniques for Training GANs

We present a variety of new architectural features and training procedures that we apply to the generative adversarial networks (GANs) framework. We focus on two applications of GANs: semi-supervised learning, and the generation of images that humans find

arxiv.org

 

Pix2Pix (PatchGAN) (2016)

논문 제목: Image-to-Image Translation with Conditional Adversarial Networks

https://arxiv.org/abs/1611.07004

 

Image-to-Image Translation with Conditional Adversarial Networks

We investigate conditional adversarial networks as a general-purpose solution to image-to-image translation problems. These networks not only learn the mapping from input image to output image, but also learn a loss function to train this mapping. This mak

arxiv.org

 

SRGAN (2016)

논문 제목: Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network

https://arxiv.org/abs/1609.04802

 

Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network

Despite the breakthroughs in accuracy and speed of single image super-resolution using faster and deeper convolutional neural networks, one central problem remains largely unsolved: how do we recover the finer texture details when we super-resolve at large

arxiv.org

 

└ Perceptual Loss (2016)

논문 제목: Perceptual Losses for Real-Time Style Transfer and Super-Resolution

https://arxiv.org/abs/1603.08155

 

Perceptual Losses for Real-Time Style Transfer and Super-Resolution

We consider image transformation problems, where an input image is transformed into an output image. Recent methods for such problems typically train feed-forward convolutional neural networks using a \emph{per-pixel} loss between the output and ground-tru

arxiv.org

 

└ StyleTr2 (2021)

논문 제목: Image Style Transfer with Transformers

https://arxiv.org/abs/2105.14576

 

StyTr$^2$: Image Style Transfer with Transformers

The goal of image style transfer is to render an image with artistic features guided by a style reference while maintaining the original content. Owing to the locality in convolutional neural networks (CNNs), extracting and maintaining the global informati

arxiv.org

 

EBGAN (2016)

논문 제목: Energy-based Generative Adversarial Network

https://arxiv.org/abs/1609.03126

 

Energy-based Generative Adversarial Network

We introduce the "Energy-based Generative Adversarial Network" model (EBGAN) which views the discriminator as an energy function that attributes low energies to the regions near the data manifold and higher energies to other regions. Similar to the probabi

arxiv.org

 

WGAN (2017)

논문 제목: Wasserstein GAN

https://arxiv.org/abs/1701.07875

 

Wasserstein GAN

We introduce a new algorithm named WGAN, an alternative to traditional GAN training. In this new model, we show that we can improve the stability of learning, get rid of problems like mode collapse, and provide meaningful learning curves useful for debuggi

arxiv.org

 

CycleGAN (2017)

논문 제목: Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks

https://arxiv.org/abs/1703.10593

 

Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks

Image-to-image translation is a class of vision and graphics problems where the goal is to learn the mapping between an input image and an output image using a training set of aligned image pairs. However, for many tasks, paired training data will not be a

arxiv.org

 

PGGAN (ProGAN) (2017)

논문 제목: Progressive Growing of GANs for Improved Quality, Stability, and Variation

https://arxiv.org/abs/1710.10196

 

Progressive Growing of GANs for Improved Quality, Stability, and Variation

We describe a new training methodology for generative adversarial networks. The key idea is to grow both the generator and discriminator progressively: starting from a low resolution, we add new layers that model increasingly fine details as training progr

arxiv.org

 

UNIT (2017)

논문 제목: Unsupervised Image-to-Image Translation Networks

https://arxiv.org/abs/1703.00848

 

Unsupervised Image-to-Image Translation Networks

Unsupervised image-to-image translation aims at learning a joint distribution of images in different domains by using images from the marginal distributions in individual domains. Since there exists an infinite set of joint distributions that can arrive th

arxiv.org

 

StyleGAN (2018)

논문 제목: A Style-Based Generator Architecture for Generative Adversarial Networks

https://arxiv.org/abs/1812.04948

 

A Style-Based Generator Architecture for Generative Adversarial Networks

We propose an alternative generator architecture for generative adversarial networks, borrowing from style transfer literature. The new architecture leads to an automatically learned, unsupervised separation of high-level attributes (e.g., pose and identit

arxiv.org

 

└ AdalN (2017)

논문 제목: Arbitrary Style Transfer in Real-time with Adaptive Instance Normalization

https://arxiv.org/abs/1703.06868

 

Arbitrary Style Transfer in Real-time with Adaptive Instance Normalization

Gatys et al. recently introduced a neural algorithm that renders a content image in the style of another image, achieving so-called style transfer. However, their framework requires a slow iterative optimization process, which limits its practical applicat

arxiv.org

 

MUNIT (2018)

논문 제목: Multimodal Unsupervised Image-to-Image Translation

https://arxiv.org/abs/1804.04732

 

Multimodal Unsupervised Image-to-Image Translation

Unsupervised image-to-image translation is an important and challenging problem in computer vision. Given an image in the source domain, the goal is to learn the conditional distribution of corresponding images in the target domain, without seeing any pair

arxiv.org

 

FUNIT (2018)

논문 제목: Few-Shot Unsupervised Image-to-Image Translation

https://arxiv.org/abs/1905.01723

 

Few-Shot Unsupervised Image-to-Image Translation

Unsupervised image-to-image translation methods learn to map images in a given class to an analogous image in a different class, drawing on unstructured (non-registered) datasets of images. While remarkably successful, current methods require access to man

arxiv.org

 

StyleGAN2 (2019)

논문 제목: Analyzing and Improving the Image Quality of StyleGAN

https://arxiv.org/abs/1912.04958

 

Analyzing and Improving the Image Quality of StyleGAN

The style-based GAN architecture (StyleGAN) yields state-of-the-art results in data-driven unconditional generative image modeling. We expose and analyze several of its characteristic artifacts, and propose changes in both model architecture and training m

arxiv.org

 

MSGGAN (2019)

논문 제목: MSG-GAN: Multi-Scale Gradients for Generative Adversarial Networks

https://arxiv.org/abs/1903.06048

 

MSG-GAN: Multi-Scale Gradients for Generative Adversarial Networks

While Generative Adversarial Networks (GANs) have seen huge successes in image synthesis tasks, they are notoriously difficult to adapt to different datasets, in part due to instability during training and sensitivity to hyperparameters. One commonly accep

arxiv.org

 

Non-saturating GAN (2020)

논문 제목: Non-saturating GAN training as divergence minimization

https://arxiv.org/abs/2010.08029

 

Non-saturating GAN training as divergence minimization

Non-saturating generative adversarial network (GAN) training is widely used and has continued to obtain groundbreaking results. However so far this approach has lacked strong theoretical justification, in contrast to alternatives such as f-GANs and Wassers

arxiv.org

 

VQ-GAN (2020)

논문 제목: Taming Transformers for High-Resolution Image Synthesis

https://arxiv.org/abs/2012.09841

 

Taming Transformers for High-Resolution Image Synthesis

Designed to learn long-range interactions on sequential data, transformers continue to show state-of-the-art results on a wide variety of tasks. In contrast to CNNs, they contain no inductive bias that prioritizes local interactions. This makes them expres

arxiv.org

 

└ VQ-VAE (2017)

논문 제목: Neural Discrete Representation Learning

https://arxiv.org/abs/1711.00937

 

Neural Discrete Representation Learning

Learning useful representations without supervision remains a key challenge in machine learning. In this paper, we propose a simple yet powerful generative model that learns such discrete representations. Our model, the Vector Quantised-Variational AutoEnc

arxiv.org

 

Drag GAN (2023)

논문 제목: Drag Your GAN: Interactive Point-based Manipulation on the Generative Image Manifold

https://arxiv.org/abs/2305.10973

 

Drag Your GAN: Interactive Point-based Manipulation on the Generative Image Manifold

Synthesizing visual content that meets users' needs often requires flexible and precise controllability of the pose, shape, expression, and layout of the generated objects. Existing approaches gain controllability of generative adversarial networks (GANs)

arxiv.org

 

Contents

포스팅 주소를 복사했습니다

이 글이 도움이 되었다면 공감 부탁드립니다.