论文标题
不变风险最小化游戏
Invariant Risk Minimization Games
论文作者
论文摘要
当在测试分布与伪造的相关性导致的训练分布不同的环境中,机器学习的标准风险最小化范式很脆弱。通过将模型集中在与结果有因果关系的特征上,对许多环境的数据培训并找到不变的预测因素可降低虚假特征的效果。在这项工作中,我们构成了这种不变的风险最小化,例如在几种环境中找到合奏游戏的NASH平衡。通过这样做,我们开发了一种简单的训练算法,该算法使用最佳响应动力学,在我们的实验中,其产生相似或更好的经验准确性,差异要比Arjovsky等人的具有挑战性的双层优化问题要低得多。 (2019)。一个关键的理论贡献表明,即使在非线性分类器和转换中,提议的游戏的NASH平衡集都等同于任何有限数量的环境的不变预测变量。结果,我们的方法还保留了Arjovsky等人所示的大量环境的概括保证。 (2019)。所提出的算法增加了成功的游戏理论机器学习算法(例如生成对抗网络)的集合。
The standard risk minimization paradigm of machine learning is brittle when operating in environments whose test distributions are different from the training distribution due to spurious correlations. Training on data from many environments and finding invariant predictors reduces the effect of spurious features by concentrating models on features that have a causal relationship with the outcome. In this work, we pose such invariant risk minimization as finding the Nash equilibrium of an ensemble game among several environments. By doing so, we develop a simple training algorithm that uses best response dynamics and, in our experiments, yields similar or better empirical accuracy with much lower variance than the challenging bi-level optimization problem of Arjovsky et al. (2019). One key theoretical contribution is showing that the set of Nash equilibria for the proposed game are equivalent to the set of invariant predictors for any finite number of environments, even with nonlinear classifiers and transformations. As a result, our method also retains the generalization guarantees to a large set of environments shown in Arjovsky et al. (2019). The proposed algorithm adds to the collection of successful game-theoretic machine learning algorithms such as generative adversarial networks.