Scaling Vision with Sparse Mixture of Experts Carlos Riquelme Google Brain Joan Puigcerver * Google Brain Basil Mustafa
–Neural Information Processing Systems
We present a Vision MoE (V -MoE), a sparse version of the Vision Transformer, that is scalable and competitive with the largest dense networks. When applied to image recognition, V -MoE matches the performance of state-of-the-art networks, while requiring as little as half of the compute at inference time.
Neural Information Processing Systems
Aug-14-2025, 07:17:40 GMT
- Country:
- Asia > Middle East
- Jordan (0.04)
- North America > Canada
- Asia > Middle East
- Industry:
- Health & Medicine (0.93)
- Technology: