Data-free Universal Adversarial Perturbation with Pseudo-semantic Prior
- Author(s)
- Lee, ChanHui
- Type
- Thesis
- Degree
- Master
- Department
- 대학원 AI대학원
- Advisor
- Son, Jeany
- Abstract
- Data-free Universal Adversarial Perturbation (UAP) is an image-agnostic adversarial attack that deceives deep neural networks using a single perturbation generated solely from random noise without relying on data priors. However, traditional data-free UAP methods often suffer from limited transferability due to the absence of semantic content in random noise. To address this issue, we propose a novel data-free universal attack method that recursively extracts pseudo-semantic priors directly from the UAPs during training to enrich the semantic content within the data-free UAP framework. Our approach effectively leverages latent semantic information within UAPs via region sampling, enabling successful input transformations—typically ineffective in traditional data-free UAP methods due to the lack of semantic cues—and significantly enhancing black-box transferabilty. Furthermore, we introduce a sample reweighting technique to mitigate potential imbalances from random sampling and transformations, emphasizing hard examples less affected by the UAPs. Comprehensive experiments on ImageNet show that our method achieves state-of-the-art performance in average fooling rate by a substantial margin, notably improves attack transferability across various CNN architectures compared to existing data-free UAP methods, and even surpasses data-dependent UAP methods. Code is available at: https://github.com/ChnanChan/PSP-UAP.
- URI
- https://scholar.gist.ac.kr/handle/local/31860
- Fulltext
- http://gist.dcollection.net/common/orgView/200000898218
- 공개 및 라이선스
-
- 파일 목록
-
Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.