Efficient Language Modeling for Low-Resource Settings with Hybrid RNN-Transformer Architectures