Prompting Adversarial Transferability via Path Flatness Attack
DOI:
https://doi.org/10.1609/aaai.v40i11.37912Abstract
Deep neural networks are susceptible to adversarial examples, which induce incorrect predictions through imperceptible perturbations. Transfer-based attacks create adversarial examples for surrogate models and transfer these examples to target models under black-box scenarios. Recent studies have established a strong correlation between the geometric properties of loss landscapes and the transferability of adversarial examples, demonstrating that flatter loss surfaces consistently yield superior transferability. However, we identify that these methods fail to account for the loss landscape flatness along the path from the current point to local minima, resulting in poor transferability. To address this, this paper constructs a novel Path Flatness Attack (PFA) method to significantly enhance the transferability of adversarial examples. Specifically, this paper proposes a novel path flatness indicator that not only evaluates the flatness in local minima regions but also explicitly quantifies the loss surface geometry along the trajectory from the current point to the minimum. Furthermore, we incorporate the path flatness indicator into the attack process, integrating penalties over low-loss points along the path while maximizing the loss function, thereby explicitly flattening the loss landscape. Extensive experiments demonstrate that PFA consistently achieves state-of-the-art attack performance across all experimental settings.Published
2026-03-14
How to Cite
Tao, Z., Peng, J., & Wang, H. (2026). Prompting Adversarial Transferability via Path Flatness Attack. Proceedings of the AAAI Conference on Artificial Intelligence, 40(11), 9511–9519. https://doi.org/10.1609/aaai.v40i11.37912
Issue
Section
AAAI Technical Track on Computer Vision VIII