论文标题

深度微分森林,对表格数据的关注很少

Deep differentiable forest with sparse attention for the tabular data

论文作者

Chen, Yingshi

论文摘要

我们提出了深层森林及其稀疏注意机制的一般建筑。可区分的森林具有树木和神经网络的优势。它的结构是一棵简单的二进制树,易于使用和理解。它具有完全的可不同性,所有变量都是可学习的参数。我们将通过基于梯度的优化方法进行训练,该方法在Deep CNN的培训中显示出强大的力量。我们发现并分析了可区分森林中的注意力机制。也就是说,每个决定仅取决于一些重要特征,而其他决定则无关紧要。注意总是很稀疏。基于此观察结果,我们通过数据感知初始化提高了它的稀疏性。我们使用属性重要性来初始化注意力重量。那么,学习的重量比随机初始化的重量要稀少。我们在一些大型表格数据集上的实验表明,可区分的森林的准确性高于GBDT,GBDT是表格数据集的最先进的算法。源代码可在https://github.com/closest-git/quantumforest上找到

We present a general architecture of deep differentiable forest and its sparse attention mechanism. The differentiable forest has the advantages of both trees and neural networks. Its structure is a simple binary tree, easy to use and understand. It has full differentiability and all variables are learnable parameters. We would train it by the gradient-based optimization method, which shows great power in the training of deep CNN. We find and analyze the attention mechanism in the differentiable forest. That is, each decision depends on only a few important features, and others are irrelevant. The attention is always sparse. Based on this observation, we improve its sparsity by data-aware initialization. We use the attribute importance to initialize the attention weight. Then the learned weight is much sparse than that from random initialization. Our experiment on some large tabular dataset shows differentiable forest has higher accuracy than GBDT, which is the state of art algorithm for tabular datasets. The source codes are available at https://github.com/closest-git/QuantumForest

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源