论文标题
用深层神经网络利用上下文信息
Exploiting Contextual Information with Deep Neural Networks
论文作者
论文摘要
上下文很重要!然而,在深度神经网络中利用上下文信息方面并没有太多研究。在大多数情况下,上下文信息的全部用法仅限于复发性神经网络。注意模型和胶囊网络是在非循环模型中引入上下文信息的两种最新方法,但是在这项工作开始后,这两种算法都是开发的。 在本论文中,我们表明可以通过两种根本不同的方式利用上下文信息:隐式和明确。在深度项目中,在上下文的使用对于识别许多微小对象的情况下非常重要,我们表明,通过精心制作卷积架构,我们可以实现最新的结果,同时也能够隐含地正确区分几乎相同的对象,但基于周围的含义不同。同时,我们表明,通过明确设计算法(源自图理论和游戏理论的动机),考虑到数据集的整个结构,我们可以实现最新的结构,从而导致了不同的主题,例如半手不足的学习和相似性学习。 据我们所知,我们是第一个集成图形理论模块的人,这些模块是为了相似性学习的问题而精心制作的,旨在考虑上下文信息,不仅要优于其他模型,而且还可以在使用较小数量的参数时获得速度提高。
Context matters! Nevertheless, there has not been much research in exploiting contextual information in deep neural networks. For most part, the entire usage of contextual information has been limited to recurrent neural networks. Attention models and capsule networks are two recent ways of introducing contextual information in non-recurrent models, however both of these algorithms have been developed after this work has started. In this thesis, we show that contextual information can be exploited in 2 fundamentally different ways: implicitly and explicitly. In the DeepScore project, where the usage of context is very important for the recognition of many tiny objects, we show that by carefully crafting convolutional architectures, we can achieve state-of-the-art results, while also being able to implicitly correctly distinguish between objects which are virtually identical, but have different meanings based on their surrounding. In parallel, we show that by explicitly designing algorithms (motivated from graph theory and game theory) that take into considerations the entire structure of the dataset, we can achieve state-of-the-art results in different topics like semi-supervised learning and similarity learning. To the best of our knowledge, we are the first to integrate graph-theoretical modules, carefully crafted for the problem of similarity learning and that are designed to consider contextual information, not only outperforming the other models, but also gaining a speed improvement while using a smaller number of parameters.