论文标题
通过自适应潜在空间抽样的有效深度表示学习
Efficient Deep Representation Learning by Adaptive Latent Space Sampling
论文作者
论文摘要
有监督的深度学习需要大量带有注释的培训样本(例如,用于分类任务的标签类,像素或体素智利的标签映射图,用于分割任务),这些图表昂贵且耗时。在训练深神经网络期间,带注释的样本以迷你批量方式馈入网络,通常将它们视为同等重要。但是,随着梯度的幅度开始消失,这些样品在训练过程中可能会变得不那么信息。同时,对于培训过程进行的其他效用或硬度的样本可能更需要进行,并且需要更多的剥削。为了应对昂贵的注释和样本信息的丢失的挑战,我们在这里提出了一个新颖的培训框架,该培训框架适应地选择了供应培训过程的信息样本。自适应选择或采样是根据生成模型构建的潜在空间中的硬度感知策略执行的。为了评估提出的培训框架,我们在三个不同的数据集上执行实验,包括用于图像分类任务的MNIST和CIFAR-10和用于生物物理模拟任务的医疗图像数据集IVU。在所有三个数据集中,所提出的框架的表现都优于随机抽样方法,该方法证明了提出的框架的有效性。
Supervised deep learning requires a large amount of training samples with annotations (e.g. label class for classification task, pixel- or voxel-wised label map for segmentation tasks), which are expensive and time-consuming to obtain. During the training of a deep neural network, the annotated samples are fed into the network in a mini-batch way, where they are often regarded of equal importance. However, some of the samples may become less informative during training, as the magnitude of the gradient start to vanish for these samples. In the meantime, other samples of higher utility or hardness may be more demanded for the training process to proceed and require more exploitation. To address the challenges of expensive annotations and loss of sample informativeness, here we propose a novel training framework which adaptively selects informative samples that are fed to the training process. The adaptive selection or sampling is performed based on a hardness-aware strategy in the latent space constructed by a generative model. To evaluate the proposed training framework, we perform experiments on three different datasets, including MNIST and CIFAR-10 for image classification task and a medical image dataset IVUS for biophysical simulation task. On all three datasets, the proposed framework outperforms a random sampling method, which demonstrates the effectiveness of proposed framework.