Goto

Collaborating Authors

 spatial attention



Controllable Text-to-Image Generation

Bowen Li, Xiaojuan Qi, Thomas Lukasiewicz, Philip Torr

Neural Information Processing Systems

Also, a word-level discriminator is proposed to providefine-grained supervisory feedback bycorrelating wordswithimageregions, facilitating training an effective generator which is able to manipulate specific visual attributes without affecting the generation of other content. Furthermore, perceptual loss is adopted to reduce the randomness involved in the image generation, andtoencourage thegenerator tomanipulate specific attributesrequired inthemodified text.





Twins: Revisiting the Design of Spatial Attention in Vision Transformers

Neural Information Processing Systems

Very recently, a variety of vision transformer architectures for dense prediction tasks have been proposed and they show that the design of spatial attention is critical to their success in these tasks. In this work, we revisit the design of the spatial attention and demonstrate that a carefully devised yet simple spatial attention mechanism performs favorably against the state-of-the-art schemes. As a result, we propose two vision transformer architectures, namely, Twins-PCPVT and Twins-SVT. Our proposed architectures are highly efficient and easy to implement, only involving matrix multiplications that are highly optimized in modern deep learning frameworks. More importantly, the proposed architectures achieve excellent performance on a wide range of visual tasks including image-level classification as well as dense detection and segmentation. The simplicity and strong performance suggest that our proposed architectures may serve as stronger backbones for many vision tasks.



A Penny for Your Thoughts: Decoding Speech from Inexpensive Brain Signals

Auster, Quentin, Shapovalenko, Kateryna, Ma, Chuang, Sun, Demaio

arXiv.org Artificial Intelligence

We explore whether neural networks can decode brain activity into speech by mapping EEG recordings to audio representations. Using EEG data recorded as subjects listened to natural speech, we train a model with a contrastive CLIP loss to align EEG-derived embeddings with embeddings from a pre-trained transformer-based speech model. Building on the state-of-the-art EEG decoder from Meta, we introduce three architectural modifications: (i) subject-specific attention layers (+0.15% WER improvement), (ii) personalized spatial attention (+0.45%), and (iii) a dual-path RNN with attention (-1.87%). Two of the three modifications improved performance, highlighting the promise of personalized architectures for brain-to-speech decoding and applications in brain-computer interfaces.



A Geometric interpretation of regularization

Neural Information Processing Systems

C HCP-Rest Resting-state Rest 1093 1200 2 HCP-Task Working Memory Task, Rest 1087 405 7 Social Mental, Random, Rest 1053 274 Relational Task, Rest 1043 232 Motor (L,R).(Hand,Foot),