论文标题
距离的感应偏见:尊重三角不平等的神经网
An Inductive Bias for Distances: Neural Nets that Respect the Triangle Inequality
论文作者
论文摘要
距离在机器学习中无处不在。它们是相似性措施,损失功能和学习目标;据说,良好的距离方法可以解决一项任务。当定义距离时,三角不平等已被证明是一个有用的约束,理论上是一种有用的限制,可以证明融合和最佳保证并在经验上是归纳偏见。尊重三角不平等的深度度量学习体系结构几乎完全依赖于潜在空间的欧几里得距离。尽管有效,但这未能建模两个广泛的亚辅助距离类别,这些距离在图形和增强学习中很常见:不对称的指标和无法嵌入欧几里得空间的指标。为了解决这些问题,我们介绍了保证满足三角形不平等的新型体系结构。我们证明了我们在$ \ mathbb {r}^n $上普遍近似规范诱导的指标,并为修改后的输入凸神经网络提供了类似的结果。我们表明,在对图形距离进行建模时,我们的体系结构的表现优于现有的度量方法,并且在多目标强化学习设置中训练数据受到限制时,与非金属方法相比具有更好的电感偏差。
Distances are pervasive in machine learning. They serve as similarity measures, loss functions, and learning targets; it is said that a good distance measure solves a task. When defining distances, the triangle inequality has proven to be a useful constraint, both theoretically--to prove convergence and optimality guarantees--and empirically--as an inductive bias. Deep metric learning architectures that respect the triangle inequality rely, almost exclusively, on Euclidean distance in the latent space. Though effective, this fails to model two broad classes of subadditive distances, common in graphs and reinforcement learning: asymmetric metrics, and metrics that cannot be embedded into Euclidean space. To address these problems, we introduce novel architectures that are guaranteed to satisfy the triangle inequality. We prove our architectures universally approximate norm-induced metrics on $\mathbb{R}^n$, and present a similar result for modified Input Convex Neural Networks. We show that our architectures outperform existing metric approaches when modeling graph distances and have a better inductive bias than non-metric approaches when training data is limited in the multi-goal reinforcement learning setting.