论文标题

长尾图像分类的复合批归一化

Compound Batch Normalization for Long-tailed Image Classification

论文作者

Cheng, Lechao, Fang, Chaowei, Zhang, Dingwen, Li, Guanbin, Huang, Gang

论文摘要

在长尾数据分布下,使用强大的培训算法(例如重新采样,重新加权和保证金调整)在学习图像分类神经网络中取得了重大进展。但是,这些方法忽略了数据不平衡对特征归一化的影响。多数类(头类别)在估计统计和仿射参数中的优势导致内部协变量在较少频繁的类别中被忽略。为了减轻这一挑战,我们提出了一种基于高斯混合物的复合批归归归式化方法。它可以更全面地建模功能空间,并降低头部类别的优势。此外,采用基于移动的平均预期最大化(EM)算法来估计多个高斯分布的统计参数。但是,EM算法对初始化很敏感,并且很容易被卡在本地最小值中,在那里,多个高斯组件继续关注多数类。为了解决此问题,我们开发了一个双路学习框架,该框架采用了类吸引的分型特征标准化来使估计的高斯分布多样化,从而使高斯组件可以更全面地适合较少频繁类的培训样本。对常用数据集的广泛实验表明,所提出的方法在长尾图像分类上优于现有方法。

Significant progress has been made in learning image classification neural networks under long-tail data distribution using robust training algorithms such as data re-sampling, re-weighting, and margin adjustment. Those methods, however, ignore the impact of data imbalance on feature normalization. The dominance of majority classes (head classes) in estimating statistics and affine parameters causes internal covariate shifts within less-frequent categories to be overlooked. To alleviate this challenge, we propose a compound batch normalization method based on a Gaussian mixture. It can model the feature space more comprehensively and reduce the dominance of head classes. In addition, a moving average-based expectation maximization (EM) algorithm is employed to estimate the statistical parameters of multiple Gaussian distributions. However, the EM algorithm is sensitive to initialization and can easily become stuck in local minima where the multiple Gaussian components continue to focus on majority classes. To tackle this issue, we developed a dual-path learning framework that employs class-aware split feature normalization to diversify the estimated Gaussian distributions, allowing the Gaussian components to fit with training samples of less-frequent classes more comprehensively. Extensive experiments on commonly used datasets demonstrated that the proposed method outperforms existing methods on long-tailed image classification.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源