A Convolution and Attention Based Encoder for Reinforcement Learning under Partial Observability

Wang, Wuhao, Chen, Zhiyong

arXiv.org Artificial Intelligence 

B. Observation History The core contribution of this work is a novel history encoder for processing historical observations, which integrates two key operations: depthwise separable convolution and multi-head attention. The background of these operations is briefly reviewed below. Depthwise separable convolution [33] is a streamlined variant of standard convolution that reduces both parameter count and computational cost. It decomposes the operation into two steps: (1) a depthwise convolution, which applies a single filter to each input channel, and (2) a pointwise convolution, which uses a 1 1 convolution to linearly combine the outputs of the depthwise stage. This factorization enables efficient extraction of spatial and cross-channel features while maintaining strong representational capacity. It has been widely adopted in lightweight neural architectures such as MobileNet [34] and is particularly well suited to real-time and resource-constrained applications. Multi-head attention [9] is a fundamental component of Transformer architectures, enabling the model to capture diverse patterns across different representation subspaces.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found