论文标题

均衡损失:长尾对象识别的梯度驱动培训

The Equalization Losses: Gradient-Driven Training for Long-tailed Object Recognition

论文作者

Tan, Jingru, Li, Bo, Lu, Xin, Yao, Yongqiang, Yu, Fengwei, He, Tong, Ouyang, Wanli

论文摘要

长尾分布在现实世界中广泛传播。由于实例的比率极小,因此尾巴类别通常显示出劣质的精度。在本文中,我们发现这种性能瓶颈主要是由不平衡的梯度引起的,该梯度可以分为两个部分:(1)积极部分,源自同一类别的样本,以及(2)负零件,由其他类别贡献。基于全面的实验,还可以观察到,累积阳性与负面因素的梯度比是衡量如何平衡类别训练的良好指标。受到这一点的启发,我们提出了一种以梯度驱动的训练机制来解决长尾问题:根据当前的累积梯度,动态地重新平衡正/负梯度,并实现平衡梯度比率的统一目标。利用简单且灵活的梯度机制,我们引入了一个新的梯度驱动损失功能,即均衡损失。我们在各种视觉任务上进行了广泛的实验,包括两阶段/单阶段长尾对象检测(LVIS),长尾图的图像分类(Imagenet-LT,Place-LT,Inaturalist)和长尾式语义分割(ADE20K)。我们的方法始终优于基线模型,证明了拟议的均衡损失的有效性和概括能力。代码将在https://github.com/modeltc/united-ception上发布。

Long-tail distribution is widely spread in real-world applications. Due to the extremely small ratio of instances, tail categories often show inferior accuracy. In this paper, we find such performance bottleneck is mainly caused by the imbalanced gradients, which can be categorized into two parts: (1) positive part, deriving from the samples of the same category, and (2) negative part, contributed by other categories. Based on comprehensive experiments, it is also observed that the gradient ratio of accumulated positives to negatives is a good indicator to measure how balanced a category is trained. Inspired by this, we come up with a gradient-driven training mechanism to tackle the long-tail problem: re-balancing the positive/negative gradients dynamically according to current accumulative gradients, with a unified goal of achieving balance gradient ratios. Taking advantage of the simple and flexible gradient mechanism, we introduce a new family of gradient-driven loss functions, namely equalization losses. We conduct extensive experiments on a wide spectrum of visual tasks, including two-stage/single-stage long-tailed object detection (LVIS), long-tailed image classification (ImageNet-LT, Places-LT, iNaturalist), and long-tailed semantic segmentation (ADE20K). Our method consistently outperforms the baseline models, demonstrating the effectiveness and generalization ability of the proposed equalization losses. Codes will be released at https://github.com/ModelTC/United-Perception.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源