MPO: An Efficient Post-Processing Framework for Mixing Diverse Preference Alignment