DeepOSets: Non-Autoregressive In-Context Learning of Supervised Learning Operators
Chiu, Shao-Ting, Hong, Junyuan, Braga-Neto, Ulisses
–arXiv.org Artificial Intelligence
We introduce DeepSets Operator Networks (DeepOSets), an efficient, non-autoregressive neural network architecture for in-context operator learning. In-context learning allows a trained machine learning model to learn from a user prompt without further training. DeepOSets adds in-context learning capabilities to Deep Operator Networks (DeepONets) by combining it with the DeepSets architecture. As the first non-autoregressive model for in-context operator learning, DeepOSets allow the user prompt to be processed in parallel, leading to significant computational savings. Here, we present the application of DeepOSets in the problem of learning supervised learning algorithms, which are operators mapping a finite-dimensional space of labeled data into an infinite-dimensional hypothesis space of prediction functions. In an empirical comparison with a popular autoregressive (transformer-based) model for in-context learning of linear regression in one and five dimensions, DeepOSets reduced the number of model weights by several orders of magnitude and required a fraction of training and inference time. Furthermore, DeepOSets proved to be less sensitive to noise, significantly outperforming the transformer model in noisy settings.
arXiv.org Artificial Intelligence
Nov-14-2024
- Country:
- Europe (0.93)
- North America > United States
- Texas > Brazos County > College Station (0.28)
- Genre:
- Research Report (0.82)
- Technology: