Towards Efficient Post-training Quantization of Pre-trained Language Models