论文标题

联合的增强的决策树具有不同的隐私

Federated Boosted Decision Trees with Differential Privacy

论文作者

Maddock, Samuel, Cormode, Graham, Wang, Tianhao, Maple, Carsten, Jha, Somesh

论文摘要

对于可以通过分布式数据进行培训的可扩展,安全和高效的隐私机器学习模型的需求很大。虽然深度学习模型通常在集中的非安全环境中取得最佳结果,但当施加隐私和通信约束时,不同的模型可以表现出色。取而代之的是,诸如XGBoost之类的基于树的方法引起了人们的高性能和易用性的关注。特别是,他们经常在表格数据上获得最新的结果。因此,最近的一些作品着重于通过加密机制(例如同型加密(HE)和安全的多方计算(MPC))来翻译诸如XGBOOST的梯度增强决策树(GBDT)模型。但是,这些并不总是提供正式的隐私保证,或者考虑各种超参数和实施设置。在这项工作中,我们在差异隐私(DP)下实施了GBDT模型。我们提出了一个通用框架,该框架捕获并扩展了差异化决策树的现有方法。我们的方法框架是针对联合设置量身定制的,我们表明,通过仔细选择技术,可以在保持强大的隐私水平的同时获得很高的效用。

There is great demand for scalable, secure, and efficient privacy-preserving machine learning models that can be trained over distributed data. While deep learning models typically achieve the best results in a centralized non-secure setting, different models can excel when privacy and communication constraints are imposed. Instead, tree-based approaches such as XGBoost have attracted much attention for their high performance and ease of use; in particular, they often achieve state-of-the-art results on tabular data. Consequently, several recent works have focused on translating Gradient Boosted Decision Tree (GBDT) models like XGBoost into federated settings, via cryptographic mechanisms such as Homomorphic Encryption (HE) and Secure Multi-Party Computation (MPC). However, these do not always provide formal privacy guarantees, or consider the full range of hyperparameters and implementation settings. In this work, we implement the GBDT model under Differential Privacy (DP). We propose a general framework that captures and extends existing approaches for differentially private decision trees. Our framework of methods is tailored to the federated setting, and we show that with a careful choice of techniques it is possible to achieve very high utility while maintaining strong levels of privacy.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源