TY - JOUR
T1 - DT-RSRGAN
T2 - An one-off domain translation generative model for real image super-resolution
AU - Zhang, Haiyu
AU - Su, Shaolin
AU - Zhu, Yu
AU - Zhang, Lingmei
AU - Yan, Qingsen
AU - Sun, Jinqiu
AU - Zhang, Yanning
N1 - Publisher Copyright:
© 2025 Elsevier Ltd
PY - 2026/1
Y1 - 2026/1
N2 - In single image super-resolution (SISR) tasks, there is inevitably a “domain gap” between synthetic and realistic datasets, which leads to performance drop accordingly. Domain translation (DT) based approaches have then emerged to narrow the discrepancy by converting data across source and target domains while maintaining semantic consistency. Currently, one-off and two-stage models constitute the main DT-SISR methods. However, due to the inability to incorporate prior knowledge of pre-trained SR networks, one-off methods often demonstrate inferior performance to the structurally complex two-stage models. To achieve both simplicity and performance gain, we propose an one-off DT-SISR model DT-RSRGAN for real-world SISR (RW-SISR). Our underlying principle is to recover LR observations via exploring vision transformer (ViT) based on self-attention (SA) mechanisms in adversarial generative models, aiming to fully explore knowledge of image internal correlation in the absence of external prior information. We then devise an image complexity (IC) loss in DT-RSRGAN, serving as a relaxed form of constraint in the absence of high-resolution (HR) training references for the one-off condition, thus suppressing artifacts that haunt GAN-based SR results. The aforementioned measures collectively facilitate the implementation of DT-RSRGAN in an one-off manner while achieving competitive performance compared to state-of-the-art (SOTA) DT-SISR solutions. Extensive experiments on multiple benchmarks validate the effectiveness and superiority of DT-RSRGAN towards RW-SISR issues.
AB - In single image super-resolution (SISR) tasks, there is inevitably a “domain gap” between synthetic and realistic datasets, which leads to performance drop accordingly. Domain translation (DT) based approaches have then emerged to narrow the discrepancy by converting data across source and target domains while maintaining semantic consistency. Currently, one-off and two-stage models constitute the main DT-SISR methods. However, due to the inability to incorporate prior knowledge of pre-trained SR networks, one-off methods often demonstrate inferior performance to the structurally complex two-stage models. To achieve both simplicity and performance gain, we propose an one-off DT-SISR model DT-RSRGAN for real-world SISR (RW-SISR). Our underlying principle is to recover LR observations via exploring vision transformer (ViT) based on self-attention (SA) mechanisms in adversarial generative models, aiming to fully explore knowledge of image internal correlation in the absence of external prior information. We then devise an image complexity (IC) loss in DT-RSRGAN, serving as a relaxed form of constraint in the absence of high-resolution (HR) training references for the one-off condition, thus suppressing artifacts that haunt GAN-based SR results. The aforementioned measures collectively facilitate the implementation of DT-RSRGAN in an one-off manner while achieving competitive performance compared to state-of-the-art (SOTA) DT-SISR solutions. Extensive experiments on multiple benchmarks validate the effectiveness and superiority of DT-RSRGAN towards RW-SISR issues.
KW - Domain translation (DT)
KW - Image complexity (IC)
KW - Real-world single image super-resolution (RW-SISR)
KW - Self-attention (SA)
KW - Vision transformer (ViT)
UR - https://www.scopus.com/pages/publications/105008775861
U2 - 10.1016/j.patcog.2025.111944
DO - 10.1016/j.patcog.2025.111944
M3 - 文章
AN - SCOPUS:105008775861
SN - 0031-3203
VL - 169
JO - Pattern Recognition
JF - Pattern Recognition
M1 - 111944
ER -