How to train your ViT for OOD Detection
Mueller, Maximilian, Hein, Matthias
–arXiv.org Artificial Intelligence
VisionTransformers have been shown to be powerful out-of-distribution detectors for ImageNet-scale settings when finetuned from publicly available checkpoints, often outperforming other model types on popular benchmarks. In this work, we investigate the impact of both the pretraining and finetuning scheme on the performance of ViTs on this task by analyzing a large pool of models. We find that the exact type of pretraining has a strong impact on which method works well and on OOD detection performance in general. We further show that certain training schemes might only be effective for a specific type of out-distribution, but not in general, and identify a best-practice training recipe. Deep neural networks have undeniably achieved remarkable success across a spectrum of realworld applications, showcasing outstanding performance. Nevertheless, they often exhibit unforeseen behaviour when confronted with unknown situations like receiving an input that is not related to the task it has been trained on.
arXiv.org Artificial Intelligence
May-21-2024
- Country:
- Europe > Germany > Baden-Württemberg > Tübingen Region > Tübingen (0.14)
- Genre:
- Research Report > New Finding (0.46)
- Technology:
- Information Technology
- Artificial Intelligence > Machine Learning
- Neural Networks (0.54)
- Performance Analysis > Accuracy (0.46)
- Statistical Learning (0.68)
- Data Science > Data Mining (0.93)
- Artificial Intelligence > Machine Learning
- Information Technology