Duplex Sequence-to-Sequence Learning for Reversible Machine Translation
–Neural Information Processing Systems
Sequence-to-sequence learning naturally has two directions. How to effectively utilize supervision signals from both directions? Existing approaches either require two separate models, or a multitask-learned model but with inferior performance. In this paper, we propose REDER (Reversible Duplex Transformer), a parameter-efficient model and apply it to machine translation.
Neural Information Processing Systems
Dec-24-2025, 18:13:22 GMT
- Technology: