Language Models are Few-Shot Learners

Neural Information Processing Systems 

Specifically, we train GPT -3, an autoregressive language model with 175 billion parameters, 10x more than any previous non-sparse language model, and test its performance in the few-shot setting.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found