Emmanuel Abbe
–Neural Information Processing Systems
Can Transformers predict new syllogisms by composing established ones? More generally, what type of targets can be learned by such models from scratch? Recent works show that Transformers can be Turing-complete in terms of expressivity, but this does not address the learnability objective. This paper puts forward the notion of globality degree of a target distribution to capture when weak learning is efficiently achievable by regular Transformers.
Neural Information Processing Systems
May-29-2025, 00:18:25 GMT
- Country:
- Europe (0.14)
- North America > United States (0.14)
- South America (0.14)
- Genre:
- Research Report > Experimental Study (1.00)
- Industry:
- Education (0.92)
- Technology: