首页 | 官方网站   微博 | 高级检索  
     

基于深度强化学习的黑盒对抗攻击算法
引用本文:李蒙,韩立新.基于深度强化学习的黑盒对抗攻击算法[J].计算机与现代化,2021,0(4):117-121.
作者姓名:李蒙  韩立新
作者单位:河海大学计算机与信息学院,江苏 南京 211100
摘    要:针对图像识别领域中的黑盒对抗攻击问题,基于强化学习中DDQN框架和Dueling网络结构提出一种黑盒对抗攻击算法。智能体通过模仿人类调整图像的方式生成对抗样本,与受攻击模型交互获得误分类结果,计算干净样本和对抗样本的结构相似性后获得奖励。攻击过程中仅获得了受攻击模型的标签输出信息。实验结果显示,攻击在CIFAR10和CIFAR100数据集上训练的4个深度神经网络模型的成功率均超过90%,生成的对抗样本质量与白盒攻击算法FGSM相近且成功率更有优势。

关 键 词:对抗样本  黑盒攻击  深度学习  强化学习  
收稿时间:2021-04-25

Black Box Adversarial Attack Algorithm Based on Deep Reinforcement Learning
LI Meng,HAN Li-xin.Black Box Adversarial Attack Algorithm Based on Deep Reinforcement Learning[J].Computer and Modernization,2021,0(4):117-121.
Authors:LI Meng  HAN Li-xin
Abstract:Aiming at the problem of black box adversarial attack in the field of image recognition, a black box adversarial attack algorithm is proposed based on the DDQN framework and Dueling network structure in reinforcement learning. The agent generates an adversarial sample by imitating human adjustment of the image, interacts with the attacked model to obtain misclassification results, and calculates the structural similarity of the clean sample and the adversarial sample to generate a reward. During the attack, only the label output information of the attacked model was obtained. The experimental results show that the success rate of attacking the four deep neural network models trained on the CIFAR10 and CIFAR100 datasets exceeds 90%. The quality of the generated adversarial samples is similar to the white box attack algorithm FGSM and the success rate is more advantageous.
Keywords:adversarial samples  black box attacks  deep learning  reinforcement learning  
本文献已被 万方数据 等数据库收录!
点击此处可从《计算机与现代化》浏览原始摘要信息
点击此处可从《计算机与现代化》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号