model on any particular supervised task). We compared with GPT-2 (345M) on the Winograd Schema Challenge
–Neural Information Processing Systems
Interesting to see how well the proposed model would do under such zero-shot setup (i.e. GPT -2 accuracy is taken from their paper. The BERT paper reports that BooksCorpus and Wikipedia contain 0.8B and 2.5B words, respectively. For our processed data, BooksCorpus and Wikipedia contain 0.75B and 2B words, respectively. The implementation is the same as word embedding, i.e., a lookup "Segment 1", and "Segment 2") and feed it to model input, which indicates the segment of input tokens.
Neural Information Processing Systems
Nov-18-2025, 15:44:06 GMT