Chunked TabPFN: Exact Training-Free In-Context Learning for Long-Context Tabular Data
Sergazinov, Renat, Yin, Shao-An
–arXiv.org Artificial Intelligence
TabPFN v2 achieves better results than tree-based models on several tabular benchmarks, which is notable since tree-based models are usually the strongest choice for tabular data. However, it cannot handle more than 10K context tokens because transformers have quadratic computation and memory costs. Unlike existing approaches that rely on context compression, such as selecting representative samples via K-nearest neighbors (KNN), we introduce a tiled-block strategy to compute attention within the TabPFN framework. This design is compatible with standard GPU setups and, to the best of our knowledge, is the first to enable TabPFN to process long contexts without any pre-processing. We demonstrate the effectiveness of our approach on the standard TabArena benchmark, with code available at chunk tabpfn.
arXiv.org Artificial Intelligence
Sep-18-2025
- Country:
- North America > United States
- Minnesota (0.04)
- Texas > Brazos County
- College Station (0.04)
- North America > United States
- Genre:
- Research Report > New Finding (0.46)
- Technology: