论文中文题名: | 基于深度学习的语音对抗样本 生成方法研究 |
姓名: | |
学号: | 21208223057 |
保密级别: | 保密(1年后开放) |
论文语种: | chi |
学科代码: | 085400 |
学科名称: | 工学 - 电子信息 |
学生类型: | 硕士 |
学位级别: | 工学硕士 |
学位年度: | 2024 |
培养单位: | 西安科技大学 |
院系: | |
专业: | |
研究方向: | 人工智能安全 |
第一导师姓名: | |
第一导师单位: | |
论文提交日期: | 2024-06-19 |
论文答辩日期: | 2024-05-31 |
论文外文题名: | Research on speech adversarial sample generation method based on deep learning |
论文中文关键词: | |
论文外文关键词: | Speech Recognition ; Adversarial Samples ; Generative Adversarial Networks ; Target Labels ; Self-attention Neural Predictor |
论文中文摘要: |
随着自动语音识别技术在自动驾驶、智能家居以及语音交互等领域的快速发展与广泛应用,其安全性与鲁棒性日益受到关注。通过精心设计的微小扰动引发的对抗样本攻击,能够迫使自动语音识别系统产生误识别,从而引发重大安全事故。为降低自动语音识别系统潜在的安全问题,研究人员通过对抗样本检测系统安全漏洞,以此提升其安全性与稳定性。因此,研究语音对抗攻击方法具有较大理论意义和实用价值。 在对抗样本生成方法中,根据攻击者对目标模型内部信息的了解程度,可分为白盒攻击和黑盒攻击两种方法。目前,白盒攻击方法面临的主要问题是难以捕获不同语音尺度间的相关性,这一局限性显著降低了攻击成功率。黑盒攻击方法虽能规避对目标模型直接访问,但其搜索过程复杂度高,并且生成过大扰动,这增加了攻击难度,也降低了对抗样本的隐蔽性。针对以上问题,本文进行如下研究: (1) 针对现有白盒攻击方法在捕捉不同语音尺度之间相关性不足的问题,提出了一种基于类别条件生成对抗网络的语音对抗攻击方法。通过目标标签映射模块将攻击目标标签转化为向量,作为条件输入到类别条件生成对抗网络中,控制样本类别的生成;设计NResidual U-block网络结构,并将其与U-Net网络相结合,能够更有效地学习不同时间尺度的语音特征,从而提高对抗样本的质量和攻击效果。在谷歌命令数据集和音乐流派数据集上的实验结果表明,与主流方法相比,本文所提语音对抗样本生成方法的攻击成功率分别提高了3.47%和5.1%,平均信噪比提升了3.2dB和1.49dB,具有良好的攻击效果和语音质量。 (2) 针对黑盒攻击过程中搜索复杂、生成扰动过大的问题,提出一种基于增强型神经预测器的黑盒语音对抗攻击方法。该方法在扰动空间中搜索最小扰动,通过自注意神经预测器指导的优化过程找到最佳扰动方向,将该方向应用于原始样本以生成对抗样本;为提高搜索效率,设计了剪枝策略,在搜索早期阶段丢弃低于阈值的样本,减少搜索次数;最后根据查询自动语音识别系统的反馈结果引入动态因子,以自适应地调整搜索步长大小,进一步加速搜索过程。为验证所提方法性能,在LibriSpeech数据集上进行实验。与主流方法相比,本文方法信噪比提升了0.8dB,样本相似度提升0.43%,查询次数平均降低7%,具有更好的攻击效果和隐蔽性。 (3) 基于上述方法,设计并开发了一个智能语音对抗攻击系统。该系统集成了本文所提攻击方法,用户可选择语音样本并自定义攻击策略与约束条件,从而实现针对特定场景的对抗攻击。该对抗攻击系统验证了对抗攻击方法的有效性,为挖掘自动语音识别系统的安全漏洞提供了有效支持。 |
论文外文摘要: |
With the rapid development and widespread application of automatic speech recognition (ASR) technology in fields such as autonomous driving, smart homes, and voice interaction, its security and robustness have become increasingly prominent concerns. Adversarial sample attacks, which are induced by carefully designed minor perturbations, can force ASR systems to produce misrecognitions, potentially leading to significant safety incidents. To mitigate the potential security issues of ASR systems, researchers have focused on detecting adversarial samples to identify system vulnerabilities and enhance their security and stability. Therefore, studying adversarial attack methods on speech recognition systems holds substantial theoretical significance and practical value. In the context of adversarial sample generation methods, attacks can be categorized into white-box attacks and black-box attacks based on the attacker’s knowledge of the target model’s internal information. Currently, the primary challenge faced by white-box attack methods is their difficulty in capturing correlations across different speech scales, which significantly reduces the attack success rate. Although black-box attack methods can avoid direct access to the target model, their search process is highly complex, and they tend to generate excessively large perturbations. This complexity increases the difficulty of the attack and reduces the concealment of the adversarial samples. To address these issues, this thesis conducts the following research: (1) To address the issue of insufficient correlation capture across different speech scales in existing white-box attack methods, a speech adversarial attack method based on category-conditional generative adversarial networks (GANs) is proposed. This method utilizes a target label mapping module to convert the attack target labels into vectors, which are then input as conditions into the category-conditional GAN to control the generation of sample categories. The designed NResidual U-block network structure, combined with the U-Net network, effectively learns speech features across different time scales, thereby improving the quality and effectiveness of the adversarial samples. Experimental results on the Google Commands dataset and the Music Genre dataset show that, compared to mainstream methods, the proposed speech adversarial sample generation method increases the attack success rates by 3.47% and 5.1%, respectively, and improves the average signal-to-noise ratio (SNR) by 3.2dB and 1.49dB. The experimental results demonstrate that the proposed method generates adversarial speech samples with excellent attack effectiveness and speech quality. (2) To tackle the challenges of high search complexity and excessive perturbation generation in black-box attacks, an enhanced neural predictor-based black-box speech adversarial attack method is proposed. This method searches for minimal perturbations within the perturbation space and uses a self-attention neural predictor to guide the optimization process, identifying the optimal perturbation direction to apply to the original samples to generate adversarial samples. To improve search efficiency, a pruning strategy is designed to discard samples below a threshold in the early search stages, reducing the number of searches. Finally, a dynamic factor is introduced based on feedback from querying the ASR system to adaptively adjust the search step size, further accelerating the search process. To validate the performance of the proposed method, experiments were conducted on the LibriSpeech dataset. Compared to mainstream methods, the proposed method improves the SNR by 0.8dB, sample similarity by 0.43%, and reduces the average number of queries by 7%. The experimental results indicate that the proposed method offers better attack effectiveness and concealment. (3) Based on the aforementioned methods, an intelligent speech adversarial attack system was designed and developed. This system integrates the proposed attack methods, allowing users to select speech samples and customize attack strategies and constraints to carry out adversarial attacks for specific scenarios. The adversarial attack system validates the practicality of the proposed methods, providing effective support for exposing security vulnerabilities in ASR systems. |
参考文献: |
[11] 刘宇宸, 宗成庆. 跨模态信息融合的端到端语音翻译[J]. 软件学报, 2022, 34(4): 1837-1849. [20] 徐东伟, 房若尘, 蒋斌, 等. 语音对抗攻击与防御方法综述[J]. 信息安全学报, 2022, 7(1): 126-144. [21] 张思思, 左信, 刘建伟. 深度学习中的对抗样本问题[J]. 计算机学报, 2019, 42(8): 1886-1904. [22] 陈晋音, 沈诗婧, 苏蒙蒙, 等. 车牌识别系统的黑盒对抗攻击[J]. 自动化学报, 2021, 47(1): 121-135. |
中图分类号: | TP391.9 |
开放日期: | 2025-06-19 |