ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks¶Star | Fork | Issue
Pipeine for Image Super-Resolution task that based on a frequently cited paper, ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks (Wang Xintao et al.), published in 2018.
In few words, image super-resolution (SR) techniques reconstruct a higher-resolution (HR) image or sequence from the observed lower-resolution (LR) images, e.g. upscaling of 720p image into 1080p.
One of the common approaches to solving this task is to use deep convolutional neural networks capable of recovering HR images from LR ones. And ESRGAN (Enhanced SRGAN) is one of them. Key points of ESRGAN:
SRResNet-based architecture with residual-in-residual blocks;
Mixture of context, perceptual, and adversarial losses. Context and perceptual losses are used for proper image upscaling, while adversarial loss pushes neural network to the natural image manifold using a discriminator network that is trained to differentiate between the super-resolved images and original photo-realistic images.
Catalyst as pipeline runner for deep learning tasks. This new and rapidly developing library can significantly reduce the amount of boilerplate code. If you are familiar with the TensorFlow ecosystem, you can think of Catalyst as Keras for PyTorch. This framework is integrated with logging systems such as the well-known TensorBoard;
Pytorch and torchvision as main frameworks for deep learning;
Albumentations and PIQ for data processing.
# step 1 - Setup environment, please check `Installation` for more info pip install git+https://github.com/leverxgroup/esrgan.git # step 2 - Load / prepare config with training details wget https://github.com/leverxgroup/esrgan/blob/master/config.yml # step 3 - train ESRGAN catalyst-dl run -C config.yml --benchmark
Some examples of work of ESRGAN model trained on DIV2K dataset:
LR (low resolution)
HR (high resolution)
The project’s GitHub repository can be found here. Bugfixes and contributions are very much appreciated!
esrgan is released under a CC-BY-NC-ND-4.0 license. See LICENSE for additional details about it.