Calibrating Scientific Foundation Models with Inference-Time Stochastic Attention
Yadav, Akash, Adebiyi, Taiwo A., Zhang, Ruda
Transformer-based scientific foundation models are increasingly deployed in high-stakes settings, but current architectures give deterministic outputs and provide limited support for calibrated predictive uncertainty. We propose Stochastic Attention, a lightweight inference-time modification that randomizes attention by replacing softmax weights with normalized multinomial samples controlled by a single concentration parameter, and produces predictive ensembles without retraining. To set this parameter, we introduce a calibration objective that matches the stochastic attention output with the target, yielding an efficient univariate post-hoc tuning problem. We evaluate this mechanism on two scientific foundation models for weather and timeseries forecasting along with an additional regression task. Across benchmarks against uncertainty-aware baselines, we find that Stochastic Attention achieves the strongest native calibration and the sharpest prediction intervals at comparable coverage, while requiring only minutes of post-hoc tuning versus days of retraining for competitive baselines.
Apr-22-2026
- Country:
- Europe > France
- Hauts-de-France > Nord > Lille (0.04)
- North America > United States
- California > Monterey County
- Monterey (0.04)
- New York > New York County
- New York City (0.04)
- Texas > Harris County
- Houston (0.14)
- California > Monterey County
- Europe > France
- Genre:
- Research Report (0.50)
- Technology: