论文标题

班级难度平衡损失用于解决类不平衡

Class-Wise Difficulty-Balanced Loss for Solving Class-Imbalance

论文作者

Sinha, Saptarshi, Ohashi, Hiroki, Nakamura, Katsuyuki

论文摘要

班级失控是现实世界数据集中的主要挑战之一,其中几个类(称为多数类)比其余的(称为少数族裔)构成的数据样本要多得多。使用此类数据集学习深层神经网络会导致通常对多数类有偏见的性能。大多数先前的作品都试图通过以各种方式为少数群体分配更多权重(例如,数据重新采样,成本敏感的学习)来解决班级失衡。但是,我们认为,可用培训数据的数量可能并不总是确定加权策略的好线索,因为即使有少数培训数据,某些少数群体也可能充分代表。超重的这些课程样本可能会导致模型的整体性能下降。我们声称,模型所感知的班级的“难度”对于确定加权更为重要。在这个角度,我们提出了一种名为“类别难度平衡损耗”的新型损失函数,或CDB损耗,该损失根据样本属于的类的难度将重量分配给每个样本的权重。请注意,随着模型的“难度”可能会随着学习的进度而变化,分配的权重动态变化。大量实验均在图像(人为地诱导的类失去平衡的MNIST,长尾cifar和Imagenet-LT)和视频(EGTEA)数据集上进行。结果表明,无论数据类型如何(即视频或图像),CDB损失始终胜过类造型数据集的最近提出的损失函数。

Class-imbalance is one of the major challenges in real world datasets, where a few classes (called majority classes) constitute much more data samples than the rest (called minority classes). Learning deep neural networks using such datasets leads to performances that are typically biased towards the majority classes. Most of the prior works try to solve class-imbalance by assigning more weights to the minority classes in various manners (e.g., data re-sampling, cost-sensitive learning). However, we argue that the number of available training data may not be always a good clue to determine the weighting strategy because some of the minority classes might be sufficiently represented even by a small number of training data. Overweighting samples of such classes can lead to drop in the model's overall performance. We claim that the 'difficulty' of a class as perceived by the model is more important to determine the weighting. In this light, we propose a novel loss function named Class-wise Difficulty-Balanced loss, or CDB loss, which dynamically distributes weights to each sample according to the difficulty of the class that the sample belongs to. Note that the assigned weights dynamically change as the 'difficulty' for the model may change with the learning progress. Extensive experiments are conducted on both image (artificially induced class-imbalanced MNIST, long-tailed CIFAR and ImageNet-LT) and video (EGTEA) datasets. The results show that CDB loss consistently outperforms the recently proposed loss functions on class-imbalanced datasets irrespective of the data type (i.e., video or image).

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源