What Happens When Small Is Made Smaller? Exploring the Impact of Compression on Small Data Pretrained Language Models