首页 | 官方网站   微博 | 高级检索  
     

双线性内卷神经网络用于眼底疾病图像分类
引用本文:杨洪刚,陈洁洁,徐梦飞.双线性内卷神经网络用于眼底疾病图像分类[J].计算机应用,2023,43(1):259-264.
作者姓名:杨洪刚  陈洁洁  徐梦飞
作者单位:湖北师范大学 计算机与信息工程学院,湖北 黄石 435002
基金项目:国家自然科学基金资助项目(61976085)
摘    要:由于眼底图像具有复杂程度高、个体差异弱、类间距离短等特点,纯卷积神经网络(CNN)和基于注意力的网络并不能在眼底疾病图像分类任务上达到令人满意的精度。因此,采用involution算子实现了注意力双线性内卷神经网络(ABINN)模型用于眼底疾病图像分类。ABINN模型的参数量仅是传统双线性卷积神经网络(BCNN)模型的11%,并提取了眼底图像的底层语义信息和空间结构信息进行二阶特征融合,是CNN和注意力方法的有效并联。此外,提出了两种基于involution算子实现注意力计算的实例化方法:基于图块的注意力子网络(AST)和基于像素的注意力子网络(ASX),这两种方法可以在CNN的基础结构内完成注意力的计算,从而使双线性子网络能在同一个架构下训练并进行特征融合。在公开眼底图像数据集OIA-ODIR上进行实验,结果显示ABINN模型的精度为85%,比通用BCNN模型提高了15.8个百分点,比TransEye模型提高了0.9个百分点。

关 键 词:眼底图像  注意力机制  involution算子  二阶特征融合  OIA-ODIR数据集
收稿时间:2021-12-08
修稿时间:2022-06-02

Bilinear involution neural network for image classification of fundus diseases
Honggang YANG,Jiejie CHEN,Mengfei XU.Bilinear involution neural network for image classification of fundus diseases[J].journal of Computer Applications,2023,43(1):259-264.
Authors:Honggang YANG  Jiejie CHEN  Mengfei XU
Affiliation:College of Computer and Information Engineering,Hubei Normal University,Huangshi Hubei 435002,China
Abstract:Due to the high complexity, weak individual differences, and short inter-class distances of fundus image features, pure Convolutional Neural Networks (CNNs) and attention based networks cannot achieve satisfactory accuracy in fundus disease image classification tasks. To this end, Attention Bilinear Involution Neural Network (ABINN) model was implemented for fundus disease image classification by using the involution operator. The parameter amount of ABINN model was only 11% of that of the traditional Bilinear Convolutional Neural Network (BCNN) model. In ABINN model, the underlying semantic information and spatial structure information of the fundus image were extracted and the second-order features of them were fused. It is an effective parallel connection between CNN and attention method. In addition, two instantiation methods for attention calculation based on involution operator, Attention Subnetwork based on PaTch (AST) and Attention Subnetwork based on PiXel (ASX), were proposed. These two methods were able to calculate attention within the CNN basic structure, thereby enabling bilinear sub-networks to be trained and fused in the same architecture. Experimental results on public fundus image dataset OIA-ODIR show that ABINN model has the accuracy of 85%, which is 15.8 percentage points higher than that of the common BCNN model and 0.9 percentage points higher than that of TransEye (Transformer Eye) model.
Keywords:fundus image  attention mechanism  involution operator  second-order feature fusion  OIA-ODIR dataset  
点击此处可从《计算机应用》浏览原始摘要信息
点击此处可从《计算机应用》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号