Neural Attention: A Novel Mechanism for Enhanced Expressive Power in Transformer Models