In AI, is bigger always better?
Artificial-intelligence systems that can churn out fluent text, such as OpenAI's ChatGPT, are the newest darlings of the technology industry. But when faced with mathematical queries that require reasoning to answer, these large language models (LLMs) often stumble. A line parallel to y 4x 6 passes through (5, 10). What is the y-coordinate of the point where this line crosses the y-axis? Although LLMs can sometimes answer these types of question correctly, they more often get them wrong. In one early test of its reasoning abilities, ChatGPT scored just 26% when faced with a sample of questions from the'MATH' data set of secondary-school-level mathematical problems1. This is to be expected: given input text, an LLM simply generates new text in accordance with statistical regularities in the words, symbols and sentences that make up the model's training data.
Mar-9-2023, 04:55:15 GMT
- Country:
- Europe
- Netherlands > North Holland
- Amsterdam (0.04)
- Switzerland > Basel-City
- Basel (0.04)
- Netherlands > North Holland
- North America
- Canada > Quebec
- Montreal (0.15)
- United States
- California > Santa Clara County
- Mountain View (0.04)
- New Mexico (0.04)
- New York (0.04)
- Oklahoma (0.04)
- Washington > King County
- California > Santa Clara County
- Canada > Quebec
- Europe
- Genre:
- Research Report > New Finding (0.47)
- Industry:
- Education > Educational Setting (0.54)
- Energy (1.00)
- Government > Regional Government (0.46)
- Information Technology > Services (0.47)
- Technology: