Sequence to Sequence Reward Modeling: Improving RLHF by Language Feedback

Authors

  • Jiayi Zhou Institute for AI, Peking University State Key Laboratory of General Artificial Intelligence, Institute for AI, Peking University
  • Jiaming Ji Institute for AI, Peking University State Key Laboratory of General Artificial Intelligence, Institute for AI, Peking University
  • Josef Dai Institute for AI, Peking University State Key Laboratory of General Artificial Intelligence, Institute for AI, Peking University
  • Yaodong Yang Institute for AI, Peking University

DOI:

https://doi.org/10.1609/aaai.v39i26.34992

Abstract

Aligning the behavior of Large language models (LLMs) with human intentions and values remains a critical challenge. Reinforcement learning from human feedback (RLHF) aligns LLMs by training a reward model (RM) on human preferences and fine-tuning the LLMs to maximize RM feedback. Despite its effectiveness and popularity, RLHF is prone to biased local optimization. It means RM fails to provide feedback that accurately aligns with human preference, causing LLMs to explore unexpected generalizations, and failing to achieve alignment objectives. To mitigate this issue, we propose a novel sequence-to-sequence (seq2seq) reward modeling method. Its key insight is that learning from language feedback rather than scalar feedback improves RLHF without additional annotations. We replaced the reward modeling target from binary maximum likelihood estimation (MLE) with sequence MLE. This method enables richer and fine-grained language feedback without additional annotations, models, or training stages. Our experiments demonstrated its effectiveness, specifically, reducing the refusal-to-response paradigm in single-turn safety dialogues and the long-response bias in text summarization tasks. We provide further analysis that seq2seq RM improves RLHF performance across 2B and 7B LLMs on 3 NLP tasks, achieving an average win rate of 76.9%. We further show that seq2seq RM can still improve the performance of RLHF under out-of-distribution prompts.

Downloads

Published

2025-04-11

How to Cite

Zhou, J., Ji, J., Dai, J., & Yang, Y. (2025). Sequence to Sequence Reward Modeling: Improving RLHF by Language Feedback. Proceedings of the AAAI Conference on Artificial Intelligence, 39(26), 27765–27773. https://doi.org/10.1609/aaai.v39i26.34992

Issue

Section

AAAI Technical Track on AI Alignment