In this paper, we present our submission to the WMT25 shared task on machine translation, for which we propose S ynergy-en h anced polic y optimization framework, named Shy . This novel two-phase training framework synergistically combines knowledge distillation and fusion via reinforcement learning. In the first phase, we introduce a multi-stage training framework that harnesses the complementary strengths of multiple state-of-the-art large language models to generate diverse, high-quality translation candidates. These candidates serve as pseudo-references to guide the supervised fine-tuning of our model, Hunyuan-7B, effectively distilling the collective knowledge of multiple expert systems into a single efficient model. In the second phase, we further refine the distilled model through Group Relative Policy Optimization, a reinforcement learning technique that employs a composite reward function. By calculating reward from multiple perspectives, our model ensures better alignment with human preferences and evaluation metrics. Extensive experiments across multiple language pairs demonstrate that our model Shy-hunyuan-MT yields substantial improvements in translation quality compared to baselines. Notably, our framework achieves competitive performance comparable to that of state-of-the-art systems while maintaining computational efficiency through knowledge distillation and fusion.
Shy-hunyuan-MT at WMT25 General Machine Translation Shared Task
Mao Zheng,Zheng Li,Yang Du,Bingxin Qu,Mingyang Song
Published 2025 in Proceedings of the Tenth Conference on Machine Translation
ABSTRACT
PUBLICATION RECORD
- Publication year
2025
- Venue
Proceedings of the Tenth Conference on Machine Translation
- Publication date
Unknown publication date
- Fields of study
Not labeled
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-18 of 18 references · Page 1 of 1
CITED BY
Showing 1-1 of 1 citing papers · Page 1 of 1