Latent Universal Task-Specific BERT
Rozental, Alon, Kelrich, Zohar, Fleischer, Daniel
This paper describes a language representation model which combines the Bidirectional Encoder Representations from Transformers (BERT) learning mechanism described in Devlin et al. (2018) with a generalization of the Universal Transformer model described in Dehghani et al. (2018). We further improve this model by adding a latent variable that represents the persona and topics of interests of the writer for each training example. We also describe a simple method to improve the usefulness of our language representation for solving problems in a specific domain at the expense of its ability to generalize to other fields. Finally, we release a pre-trained language representation model for social texts that was trained on 100 million tweets.
May-16-2019
- Country:
- Asia > Middle East
- Israel (0.14)
- North America > United States (0.28)
- Asia > Middle East
- Genre:
- Research Report (0.50)
- Technology: