Decoding Layer Saliency in Language Transformers