首页 | 官方网站   微博 | 高级检索  
     

基于边界值不变量的对抗样本检测方法
作者姓名:严飞  张铭伦  张立强
作者单位:武汉大学国家网络安全学院空天信息安全与可信计算教育部重点实验室,湖北 武汉 430072
基金项目:国家重点基础研究发展计划(“973”计划)基金资助项目(No.2014CB340601);国家自然科学基金资助项目(No.61272452)。
摘    要:目前,深度学习成为计算机领域研究与应用最广泛的技术之一,在图像识别、语音、自动驾驶、文本翻译等方面都取得良好的应用成果。但人们逐渐发现深度神经网络容易受到微小扰动图片的影响,导致分类出现错误,这类攻击手段被称为对抗样本。对抗样本的出现可能会给安全敏感的应用领域带来灾难性的后果。现有的防御手段大多需要对抗样本本身作为训练集,这种对抗样本相关的防御手段是无法应对未知对抗样本攻击的。借鉴传统软件安全中的边界检查思想,提出了一种基于边界值不变量的对抗样本检测防御方法,该方法通过拟合分布来寻找深度神经网络中的不变量,且训练集的选取与对抗样本无关。实验结果表明,在 LeNet、vgg19 模型和 Mnist、Cifar10 数据集上,与其他对抗检测方法相比,提出的方法可有效检测目前的常见对抗样本攻击,并且具有低误报率。

关 键 词:深度神经网络  边界检查  不变量  对抗样本检测  

Adversarial examples detection method based on boundary values invariants
Authors:YAN Fei  ZHANG Minglun  ZHANG Liqiang
Affiliation:Key Laboratory of Aerospace Information Security and Trusted Computing,Ministry of Education,School of Cyber Science and Engineering,Wuhan University,Wuhan 430072,China
Abstract:Nowadays,deep learning has become one of the most widely studied and applied technologies in the computer field.Deep neural networks(DNNs)have achieved greatly noticeable success in many applications such as image recognition,speech,self-driving and text translation.However,deep neural networks are vulnerable to adversarial examples that are generated by perturbing correctly classified inputs to cause DNN modes to misbehave.A boundary check method based on traditional programs by fitting the distribution to find the invariants in the deep neural network was proposed and it use the invariants to detect adversarial examples.The selection of training sets was irrelevant to adversarial examples.The experiment results show that proposed method can effectively detect the current adversarial example attacks on LeNet,vgg19 model,Mnist,Cifar10 dataset,and has a low false positive rate.
Keywords:deep neuron network  boundary checking  invariant  adversarial examples detecting
本文献已被 维普 等数据库收录!
点击此处可从《》浏览原始摘要信息
点击此处可从《》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号