论文标题

捍卫您的声音:对语音转换的对抗性攻击

Defending Your Voice: Adversarial Attack on Voice Conversion

论文作者

Huang, Chien-yu, Lin, Yist Y., Lee, Hung-yi, Lee, Lin-shan

论文摘要

近年来,语音转换方面已经取得了重大改进,这将话语的说话者特征转化为另一个说话者的说话者,而不会改变话语的语言内容。尽管如此,改进的转换技术也引起了人们对隐私和身份验证的担忧。因此,能够通过这种语音转换技术来防止自己的声音被不当使用。这就是为什么我们在本文中报告对语音转换进行对抗性攻击的首次已知尝试。我们将人类不可察觉的噪音引入一个说话者的话语中,该声音要被捍卫。鉴于这些对抗性示例,语音转换模型无法转换其他话语,从而听起来像是由辩护人制作的。对当前两个最新的零击语音转换模型进行了初步实验。据报道,白盒和黑盒情景中的客观和主观评估结果。结果表明,转换的话语的说话者特征与辩护人的说话者明显不同,而被辩护的说话者的对抗性例子与真实的话语没有区别。

Substantial improvements have been achieved in recent years in voice conversion, which converts the speaker characteristics of an utterance into those of another speaker without changing the linguistic content of the utterance. Nonetheless, the improved conversion technologies also led to concerns about privacy and authentication. It thus becomes highly desired to be able to prevent one's voice from being improperly utilized with such voice conversion technologies. This is why we report in this paper the first known attempt to perform adversarial attack on voice conversion. We introduce human imperceptible noise into the utterances of a speaker whose voice is to be defended. Given these adversarial examples, voice conversion models cannot convert other utterances so as to sound like being produced by the defended speaker. Preliminary experiments were conducted on two currently state-of-the-art zero-shot voice conversion models. Objective and subjective evaluation results in both white-box and black-box scenarios are reported. It was shown that the speaker characteristics of the converted utterances were made obviously different from those of the defended speaker, while the adversarial examples of the defended speaker are not distinguishable from the authentic utterances.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源