论文标题
可视化编码的语言信息与任务绩效之间的关系
Visualizing the Relationship Between Encoded Linguistic Information and Task Performance
论文作者
论文摘要
探测很受欢迎,可以分析是否可以通过训练有素的深度神经模型来捕获语言信息,但是很难回答编码的语言信息的变化如何影响任务绩效。为此,我们从帕累托最优性的角度研究了编码的语言信息与任务绩效之间的动态关系。它的关键思想是获得一组在两个目标方面都是帕累托最佳的模型。从这个角度来看,我们提出了一种方法来通过将其形式化为多目标优化问题来优化帕累托最佳模型。我们对两个流行的NLP任务进行实验,即机器翻译和语言建模,并研究几种语言信息和任务表现之间的关系。实验结果表明,所提出的方法比基线方法更好。我们的经验发现表明,某些句法信息有助于NLP任务,而编码更多的句法信息并不一定会带来更好的性能,因为模型体系结构也是一个重要因素。
Probing is popular to analyze whether linguistic information can be captured by a well-trained deep neural model, but it is hard to answer how the change of the encoded linguistic information will affect task performance. To this end, we study the dynamic relationship between the encoded linguistic information and task performance from the viewpoint of Pareto Optimality. Its key idea is to obtain a set of models which are Pareto-optimal in terms of both objectives. From this viewpoint, we propose a method to optimize the Pareto-optimal models by formalizing it as a multi-objective optimization problem. We conduct experiments on two popular NLP tasks, i.e., machine translation and language modeling, and investigate the relationship between several kinds of linguistic information and task performances. Experimental results demonstrate that the proposed method is better than a baseline method. Our empirical findings suggest that some syntactic information is helpful for NLP tasks whereas encoding more syntactic information does not necessarily lead to better performance, because the model architecture is also an important factor.