On Bridging the Gap between Mean Field and Finite Width in Deep Random Neural Networks with Batch Normalization