Visualizing Attention in Transformer-Based Language Representation Models