A Convolution and Attention Based Encoder for Reinforcement Learning under Partial Observability
–arXiv.org Artificial Intelligence
B. Observation History The core contribution of this work is a novel history encoder for processing historical observations, which integrates two key operations: depthwise separable convolution and multi-head attention. The background of these operations is briefly reviewed below. Depthwise separable convolution [33] is a streamlined variant of standard convolution that reduces both parameter count and computational cost. It decomposes the operation into two steps: (1) a depthwise convolution, which applies a single filter to each input channel, and (2) a pointwise convolution, which uses a 1 1 convolution to linearly combine the outputs of the depthwise stage. This factorization enables efficient extraction of spatial and cross-channel features while maintaining strong representational capacity. It has been widely adopted in lightweight neural architectures such as MobileNet [34] and is particularly well suited to real-time and resource-constrained applications. Multi-head attention [9] is a fundamental component of Transformer architectures, enabling the model to capture diverse patterns across different representation subspaces.
arXiv.org Artificial Intelligence
Sep-16-2025
- Country:
- Europe
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Serbia > Central Serbia
- Belgrade (0.04)
- Ireland > Leinster
- North America
- Canada (0.04)
- United States
- Massachusetts > Middlesex County
- Cambridge (0.04)
- Virginia > Arlington County
- Arlington (0.04)
- Massachusetts > Middlesex County
- Oceania > Australia
- New South Wales > Callaghan (0.04)
- Europe
- Genre:
- Research Report > New Finding (0.93)
- Industry:
- Health & Medicine (0.46)
- Technology: