论文标题
具有元特征学习的混合卷积神经网络,用于无线胶囊内窥镜图像中异常检测
A Hybrid Convolutional Neural Network with Meta Feature Learning for Abnormality Detection in Wireless Capsule Endoscopy Images
论文作者
论文摘要
无线胶囊内窥镜检查是检查胃肠道的最先进的非侵入性方法之一。一种用于检测胃肠道异常(例如息肉,出血,炎症等)的智能计算机辅助诊断系统在无线胶囊内窥镜图像分析中非常紧张。异常的形状,大小,颜色和纹理有很大不同,有些在视觉上与正常区域相似。由于类内部的变化,这在设计二进制分类器方面构成了挑战。在这项研究中,提出了一个混合卷积神经网络,用于使用各种卷积操作从无线胶囊内窥镜图像中提取丰富有意义的特征。它由三个平行的卷积神经网络组成,每个神经网络具有独特的特征学习能力。第一个网络利用深度可分离的卷积,而第二个网络采用余弦归一化的卷积操作。在第三个网络中引入了一种新颖的元功能提取机制,以从第一个和第二个网络及其自身上一层生成的特征中汲取的统计信息中提取模式。网络三重奏有效地处理类内方差,并有效地检测到胃肠道异常。拟议的混合卷积神经网络模型对两个广泛使用的公开数据集进行了训练和测试。测试结果表明,所提出的模型在KID和Kvasir-Capsule数据集上分别优于97 \%和98 \%分类精度的六种最先进方法。交叉数据集评估结果还证明了所提出的模型的概括性能。
Wireless Capsule Endoscopy is one of the most advanced non-invasive methods for the examination of gastrointestinal tracts. An intelligent computer-aided diagnostic system for detecting gastrointestinal abnormalities like polyp, bleeding, inflammation, etc. is highly exigent in wireless capsule endoscopy image analysis. Abnormalities greatly differ in their shape, size, color, and texture, and some appear to be visually similar to normal regions. This poses a challenge in designing a binary classifier due to intra-class variations. In this study, a hybrid convolutional neural network is proposed for abnormality detection that extracts a rich pool of meaningful features from wireless capsule endoscopy images using a variety of convolution operations. It consists of three parallel convolutional neural networks, each with a distinctive feature learning capability. The first network utilizes depthwise separable convolution, while the second employs cosine normalized convolution operation. A novel meta-feature extraction mechanism is introduced in the third network, to extract patterns from the statistical information drawn over the features generated from the first and second networks and its own previous layer. The network trio effectively handles intra-class variance and efficiently detects gastrointestinal abnormalities. The proposed hybrid convolutional neural network model is trained and tested on two widely used publicly available datasets. The test results demonstrate that the proposed model outperforms six state-of-the-art methods with 97\% and 98\% classification accuracy on KID and Kvasir-Capsule datasets respectively. Cross dataset evaluation results also demonstrate the generalization performance of the proposed model.