Goto

Collaborating Authors

 Wenhui Wang



Unified Language Model Pre-training for Natural Language Understanding and Generation

Neural Information Processing Systems

LM) that can be fine-tuned for both natural language understanding and generation tasks. The model is pre-trained using three types of language modeling tasks: unidirectional, bidirectional, and sequence-to-sequence prediction. The unified modeling is achieved by employing a shared Transformer network and utilizing specific self-attention masks to control what context the prediction conditions on.