论文标题
增强子人群之间的分布稳定性
Enhancing Distributional Stability among Sub-populations
论文作者
论文摘要
提高分布转移下机器学习算法的稳定性是分布外(OOD)泛化问题的核心。不变学习的最新作品源自因果学习,在多个培训环境中追求严格的不变性。尽管在直觉上合理,但对环境的可用性和质量进行了牢固的假设,以学习严格的不变性属性。在这项工作中,我们提出了减轻此类限制的``分配稳定性''概念。它量化了子人群之间的预测机制的稳定性,降低到规定的规模。基于此,我们提出了可学习性假设,并提出了在分布下的概括性误差,并通过理论启发了分布的启发。 W.R.T.预测机制($ y | x $ shifts)与我们的直觉相一致,并验证我们的算法的有效性。
Enhancing the stability of machine learning algorithms under distributional shifts is at the heart of the Out-of-Distribution (OOD) Generalization problem. Derived from causal learning, recent works of invariant learning pursue strict invariance with multiple training environments. Although intuitively reasonable, strong assumptions on the availability and quality of environments are made to learn the strict invariance property. In this work, we come up with the ``distributional stability" notion to mitigate such limitations. It quantifies the stability of prediction mechanisms among sub-populations down to a prescribed scale. Based on this, we propose the learnability assumption and derive the generalization error bound under distribution shifts. Inspired by theoretical analyses, we propose our novel stable risk minimization (SRM) algorithm to enhance the model's stability w.r.t. shifts in prediction mechanisms ($Y|X$-shifts). Experimental results are consistent with our intuition and validate the effectiveness of our algorithm. The code can be found at https://github.com/LJSthu/SRM.