亲爱的研友该休息了!由于当前在线用户较少,发布求助请尽量完整地填写文献信息,科研通机器人24小时在线,伴您度过漫漫科研夜!身体可是革命的本钱,早点休息,好梦!

SRFormer: Efficient Yet Powerful Transformer Network for Single Image Super Resolution

计算机科学 变压器 人工智能 卷积神经网络 模式识别(心理学) 机器学习 计算机工程 工程类 电压 电气工程
作者
Armin Mehri,Parichehr Behjati,Darío Carpio,Ángel D. Sappa
出处
期刊:IEEE Access [Institute of Electrical and Electronics Engineers]
卷期号:11: 121457-121469 被引量:4
标识
DOI:10.1109/access.2023.3328229
摘要

Recent breakthroughs in single image super resolution have investigated the potential of deep Convolutional Neural Networks (CNNs) to improve performance. However, CNNs based models suffer from their limited fields and their inability to adapt to the input content. Recently, Transformer based models were presented, which demonstrated major performance gains in Natural Language Processing and Vision tasks while mitigating the drawbacks of CNNs. Nevertheless, Transformer computational complexity can increase quadratically for high-resolution images, and the fact that it ignores the original structures of the image by converting them to the 1D structure can make it problematic to capture the local context information and adapt it for real-time applications. In this paper, we present, SRFormer, an efficient yet powerful Transformer-based architecture, by making several key designs in the building of Transformer blocks and Transformer layers that allow us to consider the original structure of the image (i.e., 2D structure) while capturing both local and global dependencies without raising computational demands or memory consumption. We also present a Gated Multi-Layer Perceptron (MLP) Feature Fusion module to aggregate the features of different stages of Transformer blocks by focusing on inter-spatial relationships while adding minor computational costs to the network. We have conducted extensive experiments on several super-resolution benchmark datasets to evaluate our approach. SRFormer demonstrates superior performance compared to state-of-the-art methods from both Transformer and Convolutional networks, with an improvement margin of 0.1 ~ 0.53 dB . Furthermore, while SRFormer has almost the same model size, it outperforms SwinIR by 0.47% and inference time by half the time of SwinIR. The code will be available on GitHub.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
隐形曼青应助专注的思松采纳,获得10
1秒前
4秒前
10秒前
10秒前
20秒前
1分钟前
LL发布了新的文献求助10
1分钟前
科研通AI6.3应助lala采纳,获得10
2分钟前
CipherSage应助科研通管家采纳,获得10
2分钟前
2分钟前
2分钟前
4分钟前
NexusExplorer应助科研通管家采纳,获得10
4分钟前
5分钟前
彭于晏应助饱满的半青采纳,获得10
6分钟前
科研通AI2S应助科研通管家采纳,获得10
6分钟前
深情安青应助ww采纳,获得10
8分钟前
ww完成签到,获得积分20
8分钟前
8分钟前
ww发布了新的文献求助10
8分钟前
豌豆苗完成签到 ,获得积分10
8分钟前
8分钟前
爆米花应助ww采纳,获得10
8分钟前
8分钟前
二狗完成签到 ,获得积分10
8分钟前
Owen应助空城采纳,获得10
8分钟前
爆米花应助饱满的半青采纳,获得10
9分钟前
9分钟前
等等发布了新的文献求助10
9分钟前
饱满的半青完成签到 ,获得积分10
9分钟前
李健应助Morwin采纳,获得10
10分钟前
文艺沉鱼完成签到 ,获得积分10
10分钟前
zhangqian完成签到 ,获得积分10
10分钟前
田様应助科研通管家采纳,获得10
10分钟前
直率海莲完成签到 ,获得积分10
10分钟前
等等发布了新的文献求助10
11分钟前
仁爱的蜻蜓完成签到,获得积分10
11分钟前
12分钟前
12分钟前
12分钟前
高分求助中
(应助此贴封号)【重要!!请各用户(尤其是新用户)详细阅读】【科研通的精品贴汇总】 10000
Elements of Propulsion: Gas Turbines and Rockets, Second Edition 1000
卤化钙钛矿人工突触的研究 1000
Engineering for calcareous sediments : proceedings of the International Conference on Calcareous Sediments, Perth 15-18 March 1988 / edited by R.J. Jewell, D.C. Andrews 1000
Wolffs Headache and Other Head Pain 9th Edition 1000
Continuing Syntax 1000
Signals, Systems, and Signal Processing 510
热门求助领域 (近24小时)
化学 材料科学 医学 生物 纳米技术 工程类 有机化学 化学工程 生物化学 计算机科学 物理 内科学 复合材料 催化作用 物理化学 光电子学 电极 细胞生物学 基因 无机化学
热门帖子
关注 科研通微信公众号,转发送积分 6246086
求助须知:如何正确求助?哪些是违规求助? 8069601
关于积分的说明 16845447
捐赠科研通 5322785
什么是DOI,文献DOI怎么找? 2834180
邀请新用户注册赠送积分活动 1811677
关于科研通互助平台的介绍 1667430