论文标题
深度主动合奏采样用于图像分类
Deep Active Ensemble Sampling For Image Classification
论文作者
论文摘要
常规的主动学习(AL)框架旨在通过积极要求对最有用的数据点进行标签来降低数据注释的成本。但是,将AL引入饥饿的深度学习算法是一个挑战。一些提出的方法包括基于不确定性的技术,几何方法,基于不确定性和几何方法的隐式组合以及基于半/自我监督技术的框架。在本文中,我们解决了该领域的两个具体问题。首先是在AL中进行样本选择中有效的开发/勘探权衡。为此,我们介绍了基于不确定性和几何框架的最新进展的创新整合,以实现样本选择策略中有效的探索/开发权衡。为此,我们建立在汤普森采样的计算有效近似值上,并以关键变化为不确定性表示的后验估计器。我们的框架提供了两个优点:(1)准确的后估计,(2)计算开销和更高准确性之间的可调整权衡。第二个问题是需要在Deep Al中改善培训方案。为此,我们使用半/自我监督学习中的想法提出了一种独立于所使用的特定技术的一般方法。综上所述,我们的框架比最新的框架有了显着改善,结果与在同一环境下监督学习的表现相当。我们显示了框架的经验结果,以及与四个数据集上的最先进的比较性能,即MNIST,CIFAR10,CIFAR100和IMAGENET,以在两个不同的设置中建立新的基线。
Conventional active learning (AL) frameworks aim to reduce the cost of data annotation by actively requesting the labeling for the most informative data points. However, introducing AL to data hungry deep learning algorithms has been a challenge. Some proposed approaches include uncertainty-based techniques, geometric methods, implicit combination of uncertainty-based and geometric approaches, and more recently, frameworks based on semi/self supervised techniques. In this paper, we address two specific problems in this area. The first is the need for efficient exploitation/exploration trade-off in sample selection in AL. For this, we present an innovative integration of recent progress in both uncertainty-based and geometric frameworks to enable an efficient exploration/exploitation trade-off in sample selection strategy. To this end, we build on a computationally efficient approximate of Thompson sampling with key changes as a posterior estimator for uncertainty representation. Our framework provides two advantages: (1) accurate posterior estimation, and (2) tune-able trade-off between computational overhead and higher accuracy. The second problem is the need for improved training protocols in deep AL. For this, we use ideas from semi/self supervised learning to propose a general approach that is independent of the specific AL technique being used. Taken these together, our framework shows a significant improvement over the state-of-the-art, with results that are comparable to the performance of supervised-learning under the same setting. We show empirical results of our framework, and comparative performance with the state-of-the-art on four datasets, namely, MNIST, CIFAR10, CIFAR100 and ImageNet to establish a new baseline in two different settings.