On the Compression of Language Models for Code: An Empirical Study on CodeBERT