Goto

Collaborating Authors

 gpt-3





TheUnreliabilityofExplanationsinFew-shot PromptingforTextualReasoning

Neural Information Processing Systems

However, text-davinci-002 is able to benefit more substantially. We further show that explanations generated by the LLMs may not entail the models' predictions norbefactually grounded intheinput, evenonsimple tasks with extractive explanations. However, these flawed explanations can still be useful as a way to verify LLMs' predictions post-hoc.





KnowGPT: KnowledgeGraphbasedPrompTingfor LargeLanguageModels

Neural Information Processing Systems

Large Language Models (LLMs) have demonstrated remarkable capabilities in many real-world applications. Nonetheless, LLMs are often criticized for their tendencytoproduce hallucinations, wherein themodels fabricate incorrectstatements on tasks beyond their knowledge and perception.


LanguageModelsareFew-ShotLearners

Neural Information Processing Systems

Specifically, we train GPT-3, an autoregressive language model with 175 billion parameters, 10x more than any previous nonsparse language model, and test its performance in the few-shot setting. For all tasks, GPT-3 is applied without any gradient updates or fine-tuning, with tasks andfew-shot demonstrations specified purelyviatextinteraction withthemodel.


ThinkBig, TeachSmall: DoLanguageModelsDistilOccam'sRazor?

Neural Information Processing Systems

Large language models have recently shown a remarkable ability for few-shot learning, including patterns of algorithmic nature. However, it is still an open question to determine what kind of patterns these models can capture and how manyexamples theyneedintheirprompts.