REINA: Regularized Entropy Information-Based Loss for Efficient Simultaneous Speech Translation
DOI:
https://doi.org/10.1609/aaai.v40i37.40360Abstract
Simultaneous Speech Translation (SimulST) systems stream in audio while simultaneously emitting translated text or speech. Such systems face the significant challenge of balancing translation quality and latency. We introduce a strategy to optimize this tradeoff: wait for more input only if you gain information by doing so. Based on this strategy, we present Regularized Entropy INformation Adaptation (REINA), a novel loss to train an adaptive policy using an existing non-streaming translation model. We derive REINA from information theory principles and show that REINA helps push the reported Pareto frontier of the latency/quality tradeoff over prior works. Utilizing REINA, we train a SimulST model on French, Spanish and German, both from and into English. Training on only open source or synthetically generated data, we achieve state-of-the-art (SOTA) streaming results for models of comparable size. We also introduce a metric for streaming efficiency, quantitatively showing REINA improves the latency/quality trade-off by as much as 21 percent compared to prior approaches, normalized against non-streaming baseline BLEU scores.Downloads
Published
2026-03-14
How to Cite
Hirschkind, N., Liu, J., Yu, X., & Nandwana, M. K. (2026). REINA: Regularized Entropy Information-Based Loss for Efficient Simultaneous Speech Translation. Proceedings of the AAAI Conference on Artificial Intelligence, 40(37), 31005-31013. https://doi.org/10.1609/aaai.v40i37.40360
Issue
Section
AAAI Technical Track on Natural Language Processing II