Enhancing Pre-training Data Detection in LLMs Through Discriminative and Symmetric Prefix Selection

Authors

  • Kai Sun Xi'an Jiaotong University
  • Yuxin Lin Xi'an Jiaotong University
  • Bo Dong Xi'an Jiaotong University
  • Jingyao Zhang Xi'an Jiaotong-Liverpool University
  • Bin Shi Xi'an Jiaotong University

DOI:

https://doi.org/10.1609/aaai.v40i39.40593

Abstract

The rapid development of large language models (LLMs) has relied on access to high-quality, large-scale datasets, yet growing concerns around data privacy and security have spurred substantial research into pre-training data detection. While state-of-the-art (SOTA) methods such as RECALL and CON-RECALL leverage auxiliary prefixes to enhance detection performance, their dependence on individual prefixes introduces notable instability across varying prefix conditions. To address this, we first conduct a theoretical analysis to assess the impact of prefixes on existing prefix-based methods. Building on the analysis, we propose a novel prefix selection method to identify optimal prefixes. Specifically, our method derives two key criteria Discriminability and Symmetry. These criteria serve to quantify the effectiveness of prefixes in detecting pre-training data, enabling precise selection of high-performing candidate prefixes. Experiments on the WikiMIA dataset demonstrate that our method consistently improves the performance of RECALL and CON-RECALL, achieving gains of up to 21.1% in AUC scores while significantly enhancing robustness.

Downloads

Published

2026-03-14

How to Cite

Sun, K., Lin, Y., Dong, B., Zhang, J., & Shi, B. (2026). Enhancing Pre-training Data Detection in LLMs Through Discriminative and Symmetric Prefix Selection. Proceedings of the AAAI Conference on Artificial Intelligence, 40(39), 33100–33107. https://doi.org/10.1609/aaai.v40i39.40593

Issue

Section

AAAI Technical Track on Natural Language Processing IV