FT-Transformer: Resilient and Reliable Transformer with End-to-End Fault Tolerant Attention
Dai, Huangliang, Wu, Shixun, Huang, Jiajun, Jian, Zizhe, Zhu, Yue, Hu, Haiyang, Chen, Zizhong
–arXiv.org Artificial Intelligence
Transformer models rely on High-Performance Computing (HPC) resources for inference, where soft errors are inevitable in large-scale systems, making the reliability of the model particularly critical. Existing fault tolerance frameworks for Transformers are designed at the operation level without architectural optimization, leading to significant computational and memory overhead, which in turn reduces protection efficiency and limits scalability to larger models. In this paper, we implement module-level protection for Transformers by treating the operations within the attention module as a single kernel and applying end-to-end fault tolerance. This method provides unified protection across multi-step computations, while achieving comprehensive coverage of potential errors in the nonlinear computations. For linear modules, we design a strided algorithm-based fault tolerance (ABFT) that avoids inter-thread communication. Experimental results show that our end-to-end fault tolerance achieves up to 7.56x speedup over traditional methods with an average fault tolerance overhead of 13.9%.
arXiv.org Artificial Intelligence
Aug-14-2025
- Country:
- Europe > Austria
- Vienna (0.14)
- North America > United States
- California > Riverside County
- Riverside (0.05)
- Nevada > Clark County
- Las Vegas (0.04)
- New York > New York County
- New York City (0.04)
- California > Riverside County
- Europe > Austria
- Genre:
- Research Report > New Finding (0.34)
- Technology: