Learning from Guidelines: Structured Prompt Optimization for Expert Annotation Tasks
DOI:
https://doi.org/10.1609/aaai.v40i41.40812Abstract
Deep learning has significantly advanced numerous fields by training on extensive annotated datasets. However, this data-driven paradigm faces limitations such as limited adaptability and high annotation costs, particularly when precise adherence to detailed, domain-specific guidelines is required in annotation. This challenge raises a critical question: Can models effectively shift from data-driven learning to autonomously leveraging guidelines with minimal annotated examples? To address this, we propose the Guideline-Driven Prompt (GDP) optimization framework, which shifts the learning paradigm from data-driven training to guideline-driven reasoning. GDP leverages Retrieval Augmented Generation (RAG) to retrieve essential fragments from complex guidelines and synthesize them into structured, executable prompts. A tree-based optimization algorithm systematically constructs and refines these prompts, explicitly capturing the intricate logic embedded in professional guidelines through a latent pipeline structure. Empirical evaluations on four datasets ranging from diverse domains and different tasks demonstrate that GDP effectively transitions the learning process from data-intensive methods to a guideline-driven approach in tasks requiring detailed and complex guideline adherence, reducing dependence on extensive annotated datasets.Published
2026-03-14
How to Cite
Zhong, W., Li, H., Dang, T. M., Jiang, F., Ma, H., Guo, Y., … Huang, J. (2026). Learning from Guidelines: Structured Prompt Optimization for Expert Annotation Tasks. Proceedings of the AAAI Conference on Artificial Intelligence, 40(41), 35068–35075. https://doi.org/10.1609/aaai.v40i41.40812
Issue
Section
AAAI Technical Track on Natural Language Processing VI