Pooling Attention: Evaluating Pretrained Transformer Embeddings for Deception Classification