Improving Transformers using Faithful Positional Encoding
Idé, Tsuyoshi, Labaien, Jokin, Chen, Pin-Yu
–arXiv.org Artificial Intelligence
We propose a new positional encoding method for a neural network architecture called the Transformer. Unlike the standard sinusoidal positional encoding, our approach is based on solid mathematical grounds and has a guarantee of not losing information about the positional order of the input sequence. We show that the new encoding approach systematically improves the prediction performance in the time-series classification task.
arXiv.org Artificial Intelligence
May-16-2024