论文标题
有效的基于亚群的会员推理攻击
An Efficient Subpopulation-based Membership Inference Attack
论文作者
论文摘要
会员推断攻击使恶意实体可以预测在培训受害者模型期间是否使用样本。最先进的会员推理攻击已证明可以达到良好的准确性,从而构成了巨大的隐私威胁。但是,大多数SOTA攻击都需要数十个阴影模型的培训,以准确推断会员资格。这种巨大的计算成本引发了有关这些攻击对深模型的实用性的问题。在本文中,我们引入了一种根本不同的MI攻击方法,该方法避免了训练数百个影子模型的需求。简而言之,我们比较了目标样本上的受害模型输出与来自相同亚群(即语义上相似的样本)的样本,而不是将其与数百个影子模型的输出进行比较。直觉是,如果目标样本不是训练样本,则模型响应在目标样本及其亚群之间不应显着差异。如果攻击者无法获得亚群样本,我们表明只有单个生成模型才能满足要求。因此,我们达到了最先进的会员推理准确性,同时大大降低了培训计算成本。
Membership inference attacks allow a malicious entity to predict whether a sample is used during training of a victim model or not. State-of-the-art membership inference attacks have shown to achieve good accuracy which poses a great privacy threat. However, majority of SOTA attacks require training dozens to hundreds of shadow models to accurately infer membership. This huge computation cost raises questions about practicality of these attacks on deep models. In this paper, we introduce a fundamentally different MI attack approach which obviates the need to train hundreds of shadow models. Simply put, we compare the victim model output on the target sample versus the samples from the same subpopulation (i.e., semantically similar samples), instead of comparing it with the output of hundreds of shadow models. The intuition is that the model response should not be significantly different between the target sample and its subpopulation if it was not a training sample. In cases where subpopulation samples are not available to the attacker, we show that training only a single generative model can fulfill the requirement. Hence, we achieve the state-of-the-art membership inference accuracy while significantly reducing the training computation cost.