Abstract
Major efforts in data-driven image super-resolution (SR) primarily focus on expanding the receptive field of the model to better capture contextual information. However, these methods are typically implemented by stacking deeper networks or leveraging transformer-based attention mechanisms, which consequently increases model complexity. In contrast, model-driven methods based on the unfolding paradigm show promise in improving performance while effectively maintaining model compactness through sophisticated module design. Based on these insights, we propose a Structural Similarity-Inspired Unfolding (SSIU) method for efficient image SR. This method is designed through unfolding an SR optimization function constrained by structural similarity, aiming to combine the strengths of both data-driven and model-driven approaches. Our model operates progressively following the unfolding paradigm. Each iteration consists of multiple Mixed-Scale Gating Modules (MSGM) and an Efficient Sparse Attention Module (ESAM). The former implements comprehensive constraints on features, including a structural similarity constraint, while the latter aims to achieve sparse activation. In addition, we design a Mixture-of-Experts-based Feature Selector (MoE-FS) that fully utilizes multi-level feature information by combining features from different steps. Extensive experiments validate the efficacy and efficiency of our unfolding-inspired network. Our model outperforms current state-of-the-art models, boasting lower parameter counts and reduced memory consumption. Our code will be available at: https://github.com/eezkni/SSIU.
| Original language | English |
|---|---|
| Pages (from-to) | 3861-3872 |
| Number of pages | 12 |
| Journal | IEEE Transactions on Image Processing |
| Volume | 34 |
| Early online date | 18 Jun 2025 |
| DOIs | |
| Publication status | Published - 2025 |
Bibliographical note
Publisher Copyright:© 1992-2012 IEEE.
Funding
This work was supported in part by the National Natural Science Foundation of China under Grant 62201387 and Grant 62371343, in part by the Fundamental Research Funds for the Central Universities, and in part by the Interdisciplinary Frontier Research Project of Pengcheng Laboratory (PCL) under Grant 2025QYB013.
Keywords
- Image super-resolution
- Light-weight
- Paradigm unfolding
- Sparse attention