Revisiting Multilingual Data Mixtures in Language Model Pretraining