Abstract

Though many attempts have been made in blind super-resolution to restore low-resolution images with unknown and complex degradations, they are still far from addressing general real-world degraded images. In this work, we extend the powerful ESRGAN to a practical restoration application (namely, Real-ESRGAN), which is trained with pure synthetic data. Specifically, a high-order degradation modeling process is introduced to better simulate complex real-world degradations. We also consider the common ringing and overshoot artifacts in the synthesis process. In addition, we employ a U-Net discriminator with spectral normalization to increase discriminator capability and stabilize the training dynamics. Extensive comparisons have shown its superior visual performance than prior works on various real datasets. We also provide efficient implementations to synthesize training pairs on the fly.

Keywords

DiscriminatorComputer scienceRingingSynthetic dataNormalization (sociology)Artificial intelligenceTraining setProcess (computing)EmbeddingImplementationReal world dataOn the flyMachine learningSpeech recognition

Affiliated Institutions

Related Publications

Publication Info

Year
2021
Type
article
Pages
1905-1914
Citations
1196
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

1196
OpenAlex

Cite This

Xintao Wang, Liangbin Xie, Chao Dong et al. (2021). Real-ESRGAN: Training Real-World Blind Super-Resolution with Pure Synthetic Data. , 1905-1914. https://doi.org/10.1109/iccvw54120.2021.00217

Identifiers

DOI
10.1109/iccvw54120.2021.00217