Pipelined Decoder for Efficient Context-Aware Text Generation
Huang, Zixian, Niu, Chenxu, Gu, Yu, Xiao, Gengyang, Huang, Xinwei, Cheng, Gong
–arXiv.org Artificial Intelligence
As the basis of generative AI, an autoregressive model requires the generation of a new token depending on all the previously generated tokens, which brings high quality but also restricts the model to generate tokens one by one, forming a bottleneck limiting the generation speed. In this paper, we propose a new decoder architecture that efficiently generates text in parallel for context-aware generation tasks. Our proposed pipelined decoder initiates the generation of multiple subsequences simultaneously, and, at each time-step, it generates a new token for each subsequence to realize parallelism. Experiments on multiple text generation tasks, including question answering, text summarization, and keyphrase generation, show that our pipelined decoder significantly improves the generation speed without a significant loss of generation quality or additional memory consumption.
arXiv.org Artificial Intelligence
Jul-2-2025
- Country:
- Africa > Ethiopia
- Addis Ababa > Addis Ababa (0.04)
- Asia
- China
- Heilongjiang Province > Daqing (0.04)
- Jiangsu Province > Nanjing (0.04)
- Japan (0.04)
- Middle East > UAE
- Abu Dhabi Emirate > Abu Dhabi (0.04)
- China
- Europe
- Austria > Vienna (0.14)
- Belgium > Brussels-Capital Region
- Brussels (0.04)
- United Kingdom > England
- West Midlands > Birmingham (0.04)
- North America
- Canada
- British Columbia > Vancouver (0.04)
- Quebec > Montreal (0.04)
- United States
- Hawaii > Honolulu County
- Honolulu (0.04)
- Louisiana > Orleans Parish
- New Orleans (0.04)
- Ohio (0.04)
- Texas > Dallas County
- Dallas (0.04)
- Washington > King County
- Seattle (0.04)
- Hawaii > Honolulu County
- Canada
- Africa > Ethiopia
- Genre:
- Research Report (0.82)
- Technology: