论文标题
从几个示例中评估室外语言模型表现
Assessing Out-of-Domain Language Model Performance from Few Examples
论文作者
论文摘要
虽然审计的语言模型表现出令人印象深刻的概括能力,但在某些领域的转变下,它们仍然表现得不可预测。特别是,模型可能会在不适合室外测试数据的内域培训数据上学习推理过程。我们以几种方式解决了预测室外(OOD)性能的任务:给定一些目标域示例和一组具有相似训练性能的模型,我们是否可以了解这些模型在OOD测试数据上的性能?在几次示例中查看模型精度时,我们在此任务上的性能进行了基准测试,然后研究如何使用功能属性来合并模型的行为分析以更好地解决此问题。具体而言,我们探索了一组旨在揭示模型一致性与某些病理启发式方法的“因素”,这些启发式方法可能表明概括能力较差。关于文本需要,解释识别和合成分类任务,我们表明基于归因的因素可以帮助对相对模型的性能进行排名。但是,几次测试集的准确性是一个令人惊讶的基线,尤其是当系统设计人员对域移位没有深入的先验知识时。
While pretrained language models have exhibited impressive generalization capabilities, they still behave unpredictably under certain domain shifts. In particular, a model may learn a reasoning process on in-domain training data that does not hold for out-of-domain test data. We address the task of predicting out-of-domain (OOD) performance in a few-shot fashion: given a few target-domain examples and a set of models with similar training performance, can we understand how these models will perform on OOD test data? We benchmark the performance on this task when looking at model accuracy on the few-shot examples, then investigate how to incorporate analysis of the models' behavior using feature attributions to better tackle this problem. Specifically, we explore a set of "factors" designed to reveal model agreement with certain pathological heuristics that may indicate worse generalization capabilities. On textual entailment, paraphrase recognition, and a synthetic classification task, we show that attribution-based factors can help rank relative model OOD performance. However, accuracy on a few-shot test set is a surprisingly strong baseline, particularly when the system designer does not have in-depth prior knowledge about the domain shift.