论文标题

LMD:可学习的掩码网络,用于检测扬声器验证的对抗示例

LMD: A Learnable Mask Network to Detect Adversarial Examples for Speaker Verification

论文作者

Chen, Xing, Wang, Jie, Zhang, Xiao-Lei, Zhang, Wei-Qiang, Yang, Kunde

论文摘要

尽管自动扬声器验证的安全性(ASV)受到最近出现的对抗性攻击的严重威胁,但仍有一些对策来减轻威胁。但是,许多防御方法不仅需要攻击者的先验知识,而且还需要弱化的解释性。为了解决这个问题,在本文中,我们提出了一种独立于攻击者的可解释方法,称为可学习的掩码探测器(LMD),以将对抗性示例与真正的示例分开。它利用分数变化作为指标来检测对抗性示例,其中得分变化是原始音频记录的ASV分数与从其掩盖的复杂频谱图合成的转换音频之间的绝对差异。得分变化检测器的核心组成部分是通过神经网络生成掩盖的频谱图。神经网络只需要进行培训的真实例子,这使其成为一种与攻击者无关的方法。它的解释性在于训练神经网络以最大程度地减少目标ASV的得分变化,并最大程度地提高真正训练示例的蒙版光谱箱数量。它的基础是基于这样的观察结果,即掩盖了绝大多数频谱图箱没有扬声器信息,这将不可避免地会引入较大的分数变化,并为真实示例带来很小的分数变化。有12位攻击者和两个代表性ASV系统的实验结果表明,我们提出的方法的表现优于五个最先进的基准。广泛的实验结果也可能是基于检测的ASV防御能力的基准。

Although the security of automatic speaker verification (ASV) is seriously threatened by recently emerged adversarial attacks, there have been some countermeasures to alleviate the threat. However, many defense approaches not only require the prior knowledge of the attackers but also possess weak interpretability. To address this issue, in this paper, we propose an attacker-independent and interpretable method, named learnable mask detector (LMD), to separate adversarial examples from the genuine ones. It utilizes score variation as an indicator to detect adversarial examples, where the score variation is the absolute discrepancy between the ASV scores of an original audio recording and its transformed audio synthesized from its masked complex spectrogram. A core component of the score variation detector is to generate the masked spectrogram by a neural network. The neural network needs only genuine examples for training, which makes it an attacker-independent approach. Its interpretability lies that the neural network is trained to minimize the score variation of the targeted ASV, and maximize the number of the masked spectrogram bins of the genuine training examples. Its foundation is based on the observation that, masking out the vast majority of the spectrogram bins with little speaker information will inevitably introduce a large score variation to the adversarial example, and a small score variation to the genuine example. Experimental results with 12 attackers and two representative ASV systems show that our proposed method outperforms five state-of-the-art baselines. The extensive experimental results can also be a benchmark for the detection-based ASV defenses.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源