Enhancing Pre-training Data Detection in LLMs Through Discriminative and Symmetric Prefix Selection
DOI:
https://doi.org/10.1609/aaai.v40i39.40593Abstract
The rapid development of large language models (LLMs) has relied on access to high-quality, large-scale datasets, yet growing concerns around data privacy and security have spurred substantial research into pre-training data detection. While state-of-the-art (SOTA) methods such as RECALL and CON-RECALL leverage auxiliary prefixes to enhance detection performance, their dependence on individual prefixes introduces notable instability across varying prefix conditions. To address this, we first conduct a theoretical analysis to assess the impact of prefixes on existing prefix-based methods. Building on the analysis, we propose a novel prefix selection method to identify optimal prefixes. Specifically, our method derives two key criteria Discriminability and Symmetry. These criteria serve to quantify the effectiveness of prefixes in detecting pre-training data, enabling precise selection of high-performing candidate prefixes. Experiments on the WikiMIA dataset demonstrate that our method consistently improves the performance of RECALL and CON-RECALL, achieving gains of up to 21.1% in AUC scores while significantly enhancing robustness.Published
2026-03-14
How to Cite
Sun, K., Lin, Y., Dong, B., Zhang, J., & Shi, B. (2026). Enhancing Pre-training Data Detection in LLMs Through Discriminative and Symmetric Prefix Selection. Proceedings of the AAAI Conference on Artificial Intelligence, 40(39), 33100–33107. https://doi.org/10.1609/aaai.v40i39.40593
Issue
Section
AAAI Technical Track on Natural Language Processing IV