首页 | 官方网站   微博 | 高级检索  
     

基于改进全局—局部注意网络的室内场景识别方法
引用本文:徐江浪,万新军,夏振平,胡伏原.基于改进全局—局部注意网络的室内场景识别方法[J].计算机应用研究,2022,39(1):316-320.
作者姓名:徐江浪  万新军  夏振平  胡伏原
作者单位:苏州科技大学电子与信息工程学院,江苏苏州215009;苏州科技大学苏州市虚拟现实智能交互及应用技术重点实验室,江苏苏州215009,苏州科技大学电子与信息工程学院,江苏苏州215009
基金项目:国家自然科学基金资助项目(61876121,62002254);江苏省重点研发计划资助项目(BE2017663);江苏省高等教育自然科学研发项目(19KJB520054)。
摘    要:由于卷积神经网络(CNN)大多侧重于全局特征学习,忽略了包含更多细节的局部特征信息,使得室内场景识别的准确率难以提高。针对这一问题,提出了基于改进全局—局部注意网络(GLANet)的室内场景识别方法。首先,利用GLANet捕捉场景图像的全局特征和局部特征,增加图像特征中的细节信息;然后,在局部网络中引入non-local注意力模块,通过注意力图和特征图的卷积来进一步保留图像的细节特征,最后融合网络不同阶段的多种特征进行分类。通过在MIT Indoor67和SUN397数据集上的训练和验证,所提方法的识别准确率与LGN方法相比分别提高了1.98%和3.07%。实验结果表明,该算法能够有效捕获全局语义信息和精细的局部细节,显著提高了识别准确率。

关 键 词:深度学习  卷积神经网络  室内场景识别  全局特征  局部特征
收稿时间:2021/5/26 0:00:00
修稿时间:2021/12/18 0:00:00

Indoor scene recognition method based on improved global-local attention network
Xu Jianglang,Wan Xinjun,Xia Zhenping and Hu Fuyuan.Indoor scene recognition method based on improved global-local attention network[J].Application Research of Computers,2022,39(1):316-320.
Authors:Xu Jianglang  Wan Xinjun  Xia Zhenping and Hu Fuyuan
Affiliation:(School of Electronic&Information Engineering,Suzhou University of Science&Technology,Suzhou Jiangsu 215009,China;Virtual Reality Key Laboratory of Intelligent Interaction&Application Technology of Suzhou,Suzhou University of Science&Technology,Suzhou Jiangsu 215009,China)
Abstract:Because convolutional neural networks(CNN) mostly focus on global feature learning and ignore local feature information containing more details, it is difficult to improve the accuracy of indoor scene recognition. To solve this problem, this paper proposed an indoor scene recognition method based on improved global-local attention network(GLANet). Firstly, it used GlANet to capture the global and local features of the scene image to increase the detail information of the image features. Then, it introduced a non-local attention module into the local network to further preserve the detail features of the image through the convolution of attention diagram and feature graph. Finally, it fused various features of different stages of the network for classification. Through training and verification on the MIT Indoor67 and the SUN397 dataset, the recognition accuracy of the proposed method is increased by 1.98% and 3.07% respectively compared with the LGN method. Experimental results show that the algorithm can effectively capture global semantic information and fine spatial details, and significantly improves the recognition accuracy.
Keywords:deep learning  convolutional neural network  indoor scene recognition  global features  local features
本文献已被 维普 万方数据 等数据库收录!
点击此处可从《计算机应用研究》浏览原始摘要信息
点击此处可从《计算机应用研究》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号