论文标题
全球可学习的关注单图像超分辨率
Global Learnable Attention for Single Image Super-Resolution
论文作者
论文摘要
自相似性对于单图超分辨率(SISR)中非本地纹理的探索很有价值。研究人员通常认为非本地纹理的重要性与其相似性得分呈正相关。在本文中,我们出人意料地发现,在修复严重损坏的查询纹理时,一些具有低相似之处的非本地纹理与目标更接近,可以提供比高相似性更准确和更丰富的细节。在这些情况下,低相似性并不意味着较低,而是由不同的尺度或方向引起的。利用这一发现,我们提出了一个全球可学习的关注(GLA),以适应训练期间非本地纹理的相似性得分,而不仅仅是使用固定的相似性评分函数,例如DOT产品。拟议的GLA可以探索具有低相似之处但更准确的细节以修复严重损坏的纹理的非本地纹理。此外,我们建议采用对超级位置敏感的哈希(SB-LSH)作为GLA的预处理方法。使用SB-LSH,相对于图像大小,我们GLA的计算复杂性从二次线性降低到渐近线性。此外,提出的GLA可以作为有效的一般构建块整合到现有的深层SISR模型中。基于GLA,我们构建了一个深厚的可学习相似性网络(DLSN),该网络实现了不同降解类型(例如模糊和噪声)的SISR任务的最新性能。我们的代码和预培训的DLSN已上传到GitHub†以验证。
Self-similarity is valuable to the exploration of non-local textures in single image super-resolution (SISR). Researchers usually assume that the importance of non-local textures is positively related to their similarity scores. In this paper, we surprisingly found that when repairing severely damaged query textures, some non-local textures with low-similarity which are closer to the target can provide more accurate and richer details than the high-similarity ones. In these cases, low-similarity does not mean inferior but is usually caused by different scales or orientations. Utilizing this finding, we proposed a Global Learnable Attention (GLA) to adaptively modify similarity scores of non-local textures during training instead of only using a fixed similarity scoring function such as the dot product. The proposed GLA can explore non-local textures with low-similarity but more accurate details to repair severely damaged textures. Furthermore, we propose to adopt Super-Bit Locality-Sensitive Hashing (SB-LSH) as a preprocessing method for our GLA. With the SB-LSH, the computational complexity of our GLA is reduced from quadratic to asymptotic linear with respect to the image size. In addition, the proposed GLA can be integrated into existing deep SISR models as an efficient general building block. Based on the GLA, we constructed a Deep Learnable Similarity Network (DLSN), which achieves state-of-the-art performance for SISR tasks of different degradation types (e.g. blur and noise). Our code and a pre-trained DLSN have been uploaded to GitHub† for validation.