Visual Decoding and Reconstruction via EEG Embeddings with Guided Diffusion
–Neural Information Processing Systems
How to decode human vision through neural signals has attracted a long-standing interest in neuroscience and machine learning. Modern contrastive learning and generative models improved the performance of visual decoding and reconstruction based on functional Magnetic Resonance Imaging (fMRI). However, the high cost and low temporal resolution of fMRI limit their applications in brain-computer interfaces (BCIs), prompting a high need for visual decoding based on electroencephalography (EEG). In this study, we present an end-to-end EEG-based visual reconstruction zero-shot framework, consisting of a tailored brain encoder, called the Adaptive Thinking Mapper (ATM), which projects neural signals from different sources into the shared subspace as the clip embedding, and a two-stage multi-pipe EEG-to-image generation strategy. In stage one, EEG is embedded to align the high-level clip embedding, and then the prior diffusion model refines EEG embedding into image priors.
Neural Information Processing Systems
Dec-27-2025, 03:23:41 GMT
- Industry:
- Health & Medicine
- Diagnostic Medicine > Imaging (0.59)
- Health Care Technology (1.00)
- Therapeutic Area > Neurology (0.39)
- Health & Medicine
- Technology: