TY - GEN
T1 - On Generating Transferable Targeted Perturbations
AU - Naseer, Muzammal
AU - Khan, Salman
AU - Hayat, Munawar
AU - Khan, Fahad Shahbaz
AU - Porikli, Fatih
N1 - Publisher Copyright:
© 2021 IEEE
PY - 2021/10
Y1 - 2021/10
N2 - While the untargeted black-box transferability of adversarial perturbations has been extensively studied before, changing an unseen model's decisions to a specific 'targeted' class remains a challenging feat. In this paper, we propose a new generative approach for highly transferable targeted perturbations (TTP). We note that the existing methods are less suitable for this task due to their reliance on class-boundary information that changes from one model to another, thus reducing transferability. In contrast, our approach matches the perturbed image 'distribution' with that of the target class, leading to high targeted transferability rates. To this end, we propose a new objective function that not only aligns the global distributions of source and target images, but also matches the local neighbourhood structure between the two domains. Based on the proposed objective, we train a generator function that can adaptively synthesize perturbations specific to a given input. Our generative approach is independent of the source or target domain labels, while consistently performs well against state-of-the-art methods on a wide range of attack settings. As an example, we achieve 32.63% target transferability from (an adversarially weak) VGG19BN to (a strong) WideResNet on ImageNet val. set, which is 4 * higher than the previous best generative attack and 16 * better than instance-specific iterative attack. Code is available at: https://github.com/Muzammal-Naseer/TTP.
AB - While the untargeted black-box transferability of adversarial perturbations has been extensively studied before, changing an unseen model's decisions to a specific 'targeted' class remains a challenging feat. In this paper, we propose a new generative approach for highly transferable targeted perturbations (TTP). We note that the existing methods are less suitable for this task due to their reliance on class-boundary information that changes from one model to another, thus reducing transferability. In contrast, our approach matches the perturbed image 'distribution' with that of the target class, leading to high targeted transferability rates. To this end, we propose a new objective function that not only aligns the global distributions of source and target images, but also matches the local neighbourhood structure between the two domains. Based on the proposed objective, we train a generator function that can adaptively synthesize perturbations specific to a given input. Our generative approach is independent of the source or target domain labels, while consistently performs well against state-of-the-art methods on a wide range of attack settings. As an example, we achieve 32.63% target transferability from (an adversarially weak) VGG19BN to (a strong) WideResNet on ImageNet val. set, which is 4 * higher than the previous best generative attack and 16 * better than instance-specific iterative attack. Code is available at: https://github.com/Muzammal-Naseer/TTP.
UR - http://www.scopus.com/inward/record.url?scp=85127811844&partnerID=8YFLogxK
UR - https://iccv2021.thecvf.com/home
U2 - 10.1109/ICCV48922.2021.00761
DO - 10.1109/ICCV48922.2021.00761
M3 - Conference contribution
AN - SCOPUS:85127811844
SN - 9781665428132
T3 - Proceedings of the IEEE International Conference on Computer Vision
SP - 7688
EP - 7697
BT - Proceedings - 2021 IEEE/CVF International Conference on Computer Vision, ICCV 2021
A2 - Berg, Tamara
A2 - Clark, James
A2 - Matsushita, Yasuyuki
A2 - Taylor, Camillo J.
A2 - Damen, Dima
A2 - Hassner, Tal
A2 - Pal, Chris
A2 - Sato, Yoichi
PB - IEEE, Institute of Electrical and Electronics Engineers
CY - United States
T2 - 18th IEEE/CVF International Conference on Computer Vision, ICCV 2021
Y2 - 11 October 2021 through 17 October 2021
ER -