遮挡对神经网络的影响:论文推荐及研究方向
遮挡对神经网络的影响:论文推荐及研究方向
遮挡是现实世界中常见的场景,它会对神经网络的性能产生负面影响。本文推荐了十篇关于遮挡对神经网络影响的论文,涵盖了对抗样本、防御机制、数据增强等方面,希望能为相关研究提供参考。
-
'Adversarial Examples Are Not Bugs, They Are Features' by Tom B. Brown et al. (2018)
-
'Explaining and Harnessing Adversarial Examples' by Ian J. Goodfellow et al. (2015)
-
'The Limitations of Deep Learning in Adversarial Settings' by Nicholas Carlini and David Wagner (2017)
-
'Towards Evaluating the Robustness of Neural Networks' by Nicholas Carlini and David Wagner (2017)
-
'Adversarial Example Detection for Deep Neural Networks' by Kexin Pei et al. (2017)
-
'Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples' by Anish Athalye et al. (2018)
-
'The Effectiveness of Data Augmentation in Image Classification using Deep Learning' by Erik H. Wouters et al. (2018)
-
'Robustness of Neural Networks to Limited Precision' by Suyog Gupta et al. (2015)
-
'Towards Deep Learning Models Resistant to Adversarial Attacks' by Aleksander Madry et al. (2017)
-
'Adversarial Robustness Toolbox v1.0.0' by IBM Research (2018)
研究方向
- 对抗样本生成: 研究如何生成更有效、更难被防御的对抗样本。
- 防御机制: 开发更有效的防御机制来抵御对抗样本攻击。
- 鲁棒性评估: 建立更完善的评估体系来衡量神经网络的鲁棒性。
- 数据增强: 通过数据增强技术提高神经网络的鲁棒性。
- 模型设计: 研究如何设计更鲁棒的模型结构。
总结
遮挡对神经网络的影响是一个重要的研究问题,上述论文只是该领域的一个缩影。随着深度学习技术的不断发展,相信该领域的研究会更加深入和广泛,并最终推动更鲁棒、更可靠的深度学习模型的诞生。
原文地址: https://www.cveoy.top/t/topic/m1BZ 著作权归作者所有。请勿转载和采集!