Deep learning has transformed AI applications but faces critical security challenges, including adversarial attacks, data poisoning, model theft, and privacy leakage. This survey examines these vulnerabilities, detailing their mechanisms and impact on model integrity and confidentiality. Practical implementations, including adversarial examples, label flipping, and backdoor attacks, are explored alongside defenses such as adversarial training, differential privacy, and federated learning, highlighting their strengths and limitations. Advanced methods like contrastive and self-supervised learning are presented for enhancing robustness. The survey concludes with future directions, emphasizing automated defenses, zero-trust architectures, and the security challenges of large AI models. A balanced approach to performance and security is essential for developing reliable deep learning systems.
翻译:深度学习已变革人工智能应用,但面临严峻的安全挑战,包括对抗攻击、数据投毒、模型窃取和隐私泄露。本综述系统审视这些漏洞,详细阐述其作用机制及对模型完整性与机密性的影响。在探究对抗样本、标签翻转和后门攻击等实际攻击手段的同时,分析了对抗训练、差分隐私和联邦学习等防御策略,并指出其优势与局限。进一步介绍了对比学习和自监督学习等先进方法在提升模型鲁棒性方面的应用。最后展望未来研究方向,强调自动化防御、零信任架构以及大型AI模型面临的安全挑战。实现性能与安全的平衡对于开发可靠的深度学习系统至关重要。