EMNLP 2025

November 06, 2025

Suzhou, China

Would you like to see your presentation here, made available to a global audience of researchers?
Add your own presentation or have us affordably record your next conference.

Our ability to efficiently and accurately evaluate the quality of machine translation systems has been outrun by the effectiveness of current language models---which limits the potential for further improving these models on more challenging tasks like long-form and literary translation. We propose an unsupervised method to generate training data for translation evaluation over different document lengths and application domains by repeated rounds of translation between source and target languages. We evaluate evaluation systems trained on texts mechanically generated using both model rotation and language translation approaches, demonstrating improved performance over a popular translation evaluation system (xCOMET) on two different tasks: (i) scoring the quality of a given translation against a human reference and (ii) selecting which of two translations is generationally closer to an original source document.

Downloads

SlidesPaperTranscript English (automatic)

Next from EMNLP 2025

CityEQA: A Hierarchical LLM Agent on Embodied Question Answering Benchmark in City Space
technical paper

CityEQA: A Hierarchical LLM Agent on Embodied Question Answering Benchmark in City Space

EMNLP 2025

+7Chen Gao
Yingfeng Chen and 9 other authors

06 November 2025

Stay up to date with the latest Underline news!

Select topic of interest (you can select more than one)

PRESENTATIONS

  • All Presentations
  • For Librarians
  • Resource Center
  • Free Trial
Underline Science, Inc.
1216 Broadway, 2nd Floor, New York, NY 10001, USA

© 2026 Underline - All rights reserved