Hardware-aware training for large-scale and diverse deep learning inference workloads using in-memory computing-based accelerators