This repository is the code of the paper("Improving ESRGAN with an additional image quality loss" [Paper]).
I referred to the site( https://github.com/Lornatang/ESRGAN-PyTorch (ESRGAN)) for this code.
cd data/
bash download_dataset.sh
use DIV2K.
Modify the contents of the file as follows.
config.py
line 38mode="train"
change tomodel="valid"
;config.py
line 104model_path=f"results/{exp_name}/g-best.pth"
change tomodel_path=f"<YOUR-WEIGHTS-PATH>.pth"
;- Run
python validate.py
.
Modify the contents of the file as follows.
config.py
line 38mode="valid"
change tomodel="train"
;- Run
python train.py
.
If you want to load weights that you've trained before, modify the contents of the file as follows.
config.py
line 38mode="valid"
change tomodel="train"
;config.py
line 56start_p_epoch=0
change tostart_p_epoch=XXX
;config.py
line 58resume=False
change toresume=True
;config.py
line 59resume_p_weight=""
change toresume_p_weight=<YOUR-RESUME-WIGHTS-PATH>
;- Run
python train.py
our results: In the following table, the result of the project.
ESRGAN / pieESRGAN / Ground Truth
If you find a bug, create a GitHub issue, or even better, submit a pull request. Similarly, if you have questions, simply post them as GitHub issues.please.
[paper link] (https://link.springer.com/article/10.1007/s11042-022-13452-4) Published: 29 July 2022
Yoonsil Choi & Hanhoon Park
Multimedia Tools and Applications (2022)Cite this article
Perceptual image super-resolution, Generateive adversarial network, ESRGAN, Additional perceptual loss, PieAPP-based image quality assessment, Multiscale disscriminator, ReLU activation
ESRGAN is a generative adversarial network that produces visually pleasing super-resolution (SR) images with high perceptual quality from low-resolution images. However, it frequently fails to recover local details, resulting in blurry or unnatural visual artifacts. To address this problem, we propose using an additional perceptual loss (computed using the pretrained PieAPP network) for training the generator, adding skip connections to the discriminator to use a combination of features with different scales, and replacing the Leaky ReLU activation functions in the discriminator with the ReLU ones. Through ×4 SR experiments utilizing real and computer-generated image benchmark datasets, it is demonstrated that the proposed method can produce SR images with significantly higher perceptual quality than ESRGAN and other ESRGAN enhancements. Specifically, when compared to ESRGAN, the proposed method resulted in 5.95 higher DMOS values, 0.46 lower PI values, and 0.01 lower LPIPS values. The source code is accessible at https://github.com/cyun-404/PieESRGAN.
@misc{wang2018esrgan,
title={ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks},
author={Xintao Wang and Ke Yu and Shixiang Wu and Jinjin Gu and Yihao Liu and Chao Dong and Chen Change Loy and Yu Qiao and Xiaoou Tang},
year={2018},
eprint={1809.00219},
archivePrefix={arXiv},
primaryClass={cs.CV}
}