Improving Transformers using Faithful Positional Encoding

Idé, Tsuyoshi, Labaien, Jokin, Chen, Pin-Yu

arXiv.org Artificial Intelligence 

We propose a new positional encoding method for a neural network architecture called the Transformer. Unlike the standard sinusoidal positional encoding, our approach is based on solid mathematical grounds and has a guarantee of not losing information about the positional order of the input sequence. We show that the new encoding approach systematically improves the prediction performance in the time-series classification task.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found