LLaMA: Open and Efficient Foundation Language Models
Touvron, Hugo, Lavril, Thibaut, Izacard, Gautier, Martinet, Xavier, Lachaux, Marie-Anne, Lacroix, Timothée, Rozière, Baptiste, Goyal, Naman, Hambro, Eric, Azhar, Faisal, Rodriguez, Aurelien, Joulin, Armand, Grave, Edouard, Lample, Guillaume
–arXiv.org Artificial Intelligence
We introduce LLaMA, a collection of foundation language models ranging from 7B to 65B parameters. We train our models on trillions of tokens, and show that it is possible to train state-of-the-art models using publicly available datasets exclusively, without resorting to proprietary and inaccessible datasets. In particular, LLaMA-13B outperforms GPT-3 (175B) on most benchmarks, and LLaMA-65B is competitive with the best models, Chinchilla-70B and PaLM-540B. We release all our models to the research community.
arXiv.org Artificial Intelligence
Feb-27-2023
- Country:
- Europe (0.93)
- North America > United States (0.28)
- Genre:
- Personal > Interview (0.67)
- Research Report (1.00)
- Industry:
- Education > Curriculum
- Subject-Specific Education (1.00)
- Energy (0.68)
- Education > Curriculum
- Technology: