Angular Gradient Sign Method: Uncovering Vulnerabilities in Hyperbolic Networks
DOI:
https://doi.org/10.1609/aaai.v40i7.37475Abstract
Adversarial examples in neural networks have been extensively studied in Euclidean settings, but recent advances in _hyperbolic networks_ call for a reevaluation of attack strategies in non-Euclidean geometries. Existing methods such as FGSM and PGD apply perturbations without regard to the underlying hyperbolic structure, potentially leading to inefficient or geometrically inconsistent attacks. In this work, we propose a novel adversarial attack that explicitly leverages the geometric properties of hyperbolic space. Specifically, we compute the gradient of the loss function in the tangent space of hyperbolic space, decompose it into a radial (depth) component and an angular (semantic) component, and apply perturbation derived solely from the angular direction. Our method generates adversarial examples by focusing perturbations in semantically sensitive directions encoded in angular movement within the hyperbolic geometry. Empirical results on image classification, cross-modal retrieval tasks and network architectures demonstrate that our attack achieves higher fooling rates than conventional adversarial attacks, while producing high-impact perturbations with deeper insights into vulnerabilities of hyperbolic embeddings. This work highlights the importance of geometry-aware adversarial strategies in curved representation spaces and provides a principled framework for attacking hierarchical embeddings.Downloads
Published
2026-03-14
How to Cite
Jo, M., Yang, D., & Kim, T. (2026). Angular Gradient Sign Method: Uncovering Vulnerabilities in Hyperbolic Networks. Proceedings of the AAAI Conference on Artificial Intelligence, 40(7), 5566–5574. https://doi.org/10.1609/aaai.v40i7.37475
Issue
Section
AAAI Technical Track on Computer Vision IV