论文标题
跨跨性别的几个学习:带视觉声音元音的粗到伪标记
Towards Cross-Granularity Few-Shot Learning: Coarse-to-Fine Pseudo-Labeling with Visual-Semantic Meta-Embedding
论文作者
论文摘要
很少有学习的目的是在测试时间仅使用少数样本快速适应新的类别,而元学习的想法主要是针对的。但是,元学习方法实质上是在各种少数射击任务中学习的,因此仍然需要大规模的培训数据,并具有细粒度的监督以得出广义模型,从而涉及过度的注释成本。在本文中,我们将少量的分类范式推向了更具挑战性的情况,即跨粒度分类,该模型在训练过程中仅观察到粗标签,而期望在测试过程中执行细粒度分类。这项任务在很大程度上可以减轻注释成本,因为细粒度的标签通常需要特定于领域的专业知识。为了弥合跨粒度差距,我们根据图像嵌入的相似性,通过将每个粗级级别的贪婪聚类近似于细粒度的数据分布。然后,我们提出了一个元插入器,以在实例和粗大的情况下共同优化视觉和语义歧视,以获得此粗到最新的伪标记过程的良好特征空间。进行了广泛的实验和消融研究,以证明我们方法在三个代表性数据集上的有效性和鲁棒性。
Few-shot learning aims at rapidly adapting to novel categories with only a handful of samples at test time, which has been predominantly tackled with the idea of meta-learning. However, meta-learning approaches essentially learn across a variety of few-shot tasks and thus still require large-scale training data with fine-grained supervision to derive a generalized model, thereby involving prohibitive annotation cost. In this paper, we advance the few-shot classification paradigm towards a more challenging scenario, i.e., cross-granularity few-shot classification, where the model observes only coarse labels during training while is expected to perform fine-grained classification during testing. This task largely relieves the annotation cost since fine-grained labeling usually requires strong domain-specific expertise. To bridge the cross-granularity gap, we approximate the fine-grained data distribution by greedy clustering of each coarse-class into pseudo-fine-classes according to the similarity of image embeddings. We then propose a meta-embedder that jointly optimizes the visual- and semantic-discrimination, in both instance-wise and coarse class-wise, to obtain a good feature space for this coarse-to-fine pseudo-labeling process. Extensive experiments and ablation studies are conducted to demonstrate the effectiveness and robustness of our approach on three representative datasets.