期刊论文详细信息
IEEE Access
Besting the Black-Box: Barrier Zones for Adversarial Example Defense
Thanh Nguyen1  Marten Van Dijk2  Kaleel Mahmood3  Phuong Ha Nguyen4  Lam M. Nguyen5 
[1] Amazon Inc., Seattle, WA, USA;CWI Amsterdam, Amsterdam, The Netherlands;Department of Electrical and Computer Engineering, University of Connecticut, Storrs, CT, USA;EBay Inc., San Jose, CA, USA;IBM Research, Thomas J. Watson Research Center, Yorktown Heights, NY, USA;
关键词: Adversarial machine learning;    adversarial examples;    adversarial defense;    black-box attack;    security;    deep learning;   
DOI  :  10.1109/ACCESS.2021.3138966
来源: DOAJ
【 摘 要 】

Adversarial machine learning defenses have primarily been focused on mitigating static, white-box attacks. However, it remains an open question whether such defenses are robust under an adaptive black-box adversary. In this paper, we specifically focus on the black-box threat model and make the following contributions: First we develop an enhanced adaptive black-box attack which is experimentally shown to be $\geq 30\%$ more effective than the original adaptive black-box attack proposed by Papernot et al. For our second contribution, we test 10 recent defenses using our new attack and propose our own black-box defense (barrier zones). We show that our defense based on barrier zones offers significant improvements in security over state-of-the-art defenses. This improvement includes greater than 85% robust accuracy against black-box boundary attacks, transfer attacks and our new adaptive black-box attack, for the datasets we study. For completeness, we verify our claims through extensive experimentation with 10 other defenses using three adversarial models (14 different black-box attacks) on two datasets (CIFAR-10 and Fashion-MNIST).

【 授权许可】

Unknown   

  文献评价指标  
  下载次数:0次 浏览次数:4次