论文标题
深度多视图增强效果用于图像检索
Deep Multi-View Enhancement Hashing for Image Retrieval
论文作者
论文摘要
哈希是一种有效的方法,可以通过将高维特征描述符嵌入具有较低尺寸的锤击空间的相似性中,可以在大规模数据空间中最近进行邻居搜索。但是,与传统的检索方法相比,通过二进制代码进行的大规模高速检索的检索准确性有一定程度的降低。我们已经注意到,多视图方法可以很好地保留数据的各种特征。因此,我们尝试将多视图深神经网络介绍到哈希学习领域,并设计一个有效且创新的检索模型,该模型在检索性能方面取得了重大改进。在本文中,我们提出了一个有监督的多视图模型,该模型可以通过神经网络增强多视图信息。这是一种结合多视图和深度学习方法的全新哈希学习方法。所提出的方法利用有效的视图稳定性评估方法积极探索观点之间的关系,这将影响整个网络的优化方向。我们还设计了锤子空间中的多种多数据融合方法,以保持卷积和多视图的优势。为了避免在检索过程中的增强程序上过多的计算资源,我们建立了一个称为内存网络的单独结构,该结构共同参与培训。该方法在CIFAR-10,NUS范围内和MS-Coco数据集上进行系统评估,结果表明,我们的方法显着优于最新的单视图和多视图哈希方法。
Hashing is an efficient method for nearest neighbor search in large-scale data space by embedding high-dimensional feature descriptors into a similarity preserving Hamming space with a low dimension. However, large-scale high-speed retrieval through binary code has a certain degree of reduction in retrieval accuracy compared to traditional retrieval methods. We have noticed that multi-view methods can well preserve the diverse characteristics of data. Therefore, we try to introduce the multi-view deep neural network into the hash learning field, and design an efficient and innovative retrieval model, which has achieved a significant improvement in retrieval performance. In this paper, we propose a supervised multi-view hash model which can enhance the multi-view information through neural networks. This is a completely new hash learning method that combines multi-view and deep learning methods. The proposed method utilizes an effective view stability evaluation method to actively explore the relationship among views, which will affect the optimization direction of the entire network. We have also designed a variety of multi-data fusion methods in the Hamming space to preserve the advantages of both convolution and multi-view. In order to avoid excessive computing resources on the enhancement procedure during retrieval, we set up a separate structure called memory network which participates in training together. The proposed method is systematically evaluated on the CIFAR-10, NUS-WIDE and MS-COCO datasets, and the results show that our method significantly outperforms the state-of-the-art single-view and multi-view hashing methods.